Introduction
Notakey Authentication Appliance (NtkAA) is a RancherOS Linux-based virtual machine, which contains the core Notakey API software, a RADIUS proxy for transparently integrating the Notakey API into existing RADIUS-based workflows, a Single Sign-On (SSO) web service for SAML and OAuth and other standard authentication support and various plugins for integration with cloud and local services.
NtkAA supports various redundancy and load balancing schemes and fully supports zero downtime rolling updates that can be requested with single command.
Scaling is supported by underlying cluster architecture allowing additional member joining with no downtime and advanced configuration of role based replication. Groups of nodes can be delegated to perform specific tasks like mobile API backend, SSO frontend and others.
Product Versions in Appliance
This user manual’s version number will match the corresponding appliance’s version number.
Each appliance contains the following software:
- Notakey Authentication Server (NtkAs) (version: 3.1.1)
- Notakey Authentication Proxy (NtkAp) (version: 1.0.5)
- VRRP agent (VRRPa) (version: 2.1.5-1)
- Notakey SSO (NtkSSO) (version: 3.1.7)
The appliance version number, or this user manual’s version number, will not typically match the version numbers of either NtkAs or NtkAp.
Technical Summary
The appliance is a self-contained environment, with no (mandatory) external dependencies, where the Notakey software is installed during setup wizard. To avoid any potential software conflicts, the VM used for Notakey software installation should be single-purpose.
It is stripped down as much as possible - only the bare minimum system components, which are required for the software to operate, are left on the system. This keeps the potential attack surface to a minimum.
Appliance can be be provisioned with limited special purpose user for daily management tasks during install time. This user will have all necessary management privilledges for provided services, but no system service level access.
Appliance Licences
Each appliance cluster has an associated licence-type and an associated version.
Appliances with the same version, but different licences, support the same features. Feature availability is determined by the licence version. Throughout this document, features whose availability depends on the licence type, will be specially annotated.
Licence Type | Description |
---|---|
Express | Licence for deployments with under 2000 users. |
Enterprise | Licence for deployments with over 10 000 users. |
Express licence
Per user licences are issued when user onboards a mobile device and remains locked to this user until onboarded device is removed (offboarded). Contact your sales representative for more details.
Enterprise licence
User licence usage is calculated based on a fixed day rolling window, considering only users that have any authentication or onboarding activity in this window. Contact your sales representative for more details.
System Requirements
Infrastructure Requirements
- Local network for appliance virtual machines (VMs)
Notakey Appliance VM requires at least one internal network interface with single internal IP address and public internet access during setup. If using multiple VMs this should be a dedicated network for Notakey services containing only Notakey Appliance VMs and router.
- Dedicated public IP address
A single public IP address is required to be able to set up port forwarding and issue a DNS name for service. This does not apply to load balancer or cloud deployment scenarios.
- Ingress controller
In most basic use case this will be your local network router with firewall and destination network address translation (dNAT) capability. You will need to forward port tcp/443 (https) of public IP address to appliance VM internal IP address. You will also need to forward port tcp/80 (http) if using automatic Let’s Encrypt TLS certificates for domain name validation to succeed. Larger scale deplyments can use external load balancers to be able to scale service as required. It is also possible to deploy Notakey services to your Kubernetes cloud infrastructure, please contact our support for more details.
- Dedicated DNS name for service
DNS name is required to be able to assign and validate TLS certificate. The Notakey Authenticator application requires TLS for operation.
- (Optional) TLS certificate from trusted CA
If not using automatic Let’s Encrypt TLS certificate, you will need to provide TLS key and certificate file (PEM base64 encoded with certification chain) for the dedicated DNS name used for Authentication Server service.
Hypervisor Compatibility
The following table is a appliance version/hypervisor compatibility matrix:
Appliance Version | Supported Virtualization Environments |
---|---|
<= 5.* | Hyper-V (Windows Server 2012 R2 and newer) |
VMware ESXi 5.5 (and newer) | |
Oracle Virtualbox 5.2 (and newer) |
VMware Information
The following table contains information about different appliance versions, as it pertains to VMware ESXi.
Appliance Version | Virtual Hardware Version |
---|---|
<= 3.* | vmx-09 |
Hardware Requirements
The Notakey service is typically CPU bound. Most operations are reading from an in-memory store.
Minimum Requirements (Less Than 1000 Users)
For a deployment with less than 1000 users, the suggested requirements are:
Resource | Value |
---|---|
RAM | 4 GB |
vCPU | 2 |
CPU type | 2.3 GHz Intel Xeon® E5-2686 v4 (Broadwell) or newer |
Storage | 32 GB |
Suggested Requirements
For a typical deployment, with up to 300 000 users, the suggested hardware requirements are:
Resource | Value |
---|---|
RAM | 8 GB |
vCPU | 8 |
CPU type | 2.3 GHz Intel Xeon® E5-2686 v4 (Broadwell) or newer |
Storage | 256 GB |
Availability considerations
In case of 3-node deployments, if 1 node fails its healthcheck (or becomes unavailable), traffic can be routed to the other route.
NOTE: more than half nodes are required to be operational for full service. If more than half nodes are not operational, then the remaining nodes will operate in a limited capacity.
It will allow:
- new authentication requests for existing users based on previously issued API access tokens
It will not allow:
onboarding new users
changing onboarding requirements
creating new applications
editing existing applications or users
issuing new API access tokens
In addition, the partially operational nodes will not log authorization requests
in the web user interface. It will still be possible, however, to log these requests
to rsyslog
(if so configured).
General failure tolerance depends on the total number of nodes. To maintain data integrity, a quorum of servers must be available, which can be calculated with the formula Cquorum = FLOOR(Cservers / 2) + 1.
The failure tolerance can be calculated by FT = Cservers - Cquorum.
In other words - more than half of the total available servers must be available and operational.
Total servers | Quorum size | Failure tolerance |
---|---|---|
1 | 1 | 0 |
2 | 2 | 0 |
3 | 2 | 1 |
4 | 3 | 1 |
5 | 3 | 2 |
6 | 4 | 2 |
7 | 4 | 3 |
8 | 5 | 3 |
9 | 5 | 4 |
10 | 6 | 4 |
Network Connectivity
Generally only port required for normal operation is TCP/443, if TLS is terminated on appliance. A configuration with external load balancers is available too, in this case port numbers are provided depending on service. See individual service documentation below for more information.
Appliance connects to the outside word via HTTPS and requires several hosts available normal operation. Update repository and push notification service can be also run on premise, if required. Contact your support representative for more details.
Appliance Installation
Appliance install is performed by installing base RancerOS container Linux. Any version since 1.1 should be ok, but we recommend installing latest release version. After OS is installed to disk and network configured, Notakey services can be installed by notakey/cfg container image.
Download latest RancherOS installation ISO from https://github.com/rancher/os/releases.
Provision virtual machine according to your hardware requirements.
Boot virtual machine from ISO rancheros.iso.
Provision appliance with cloud-config (optional).
** Set temporary password for installer
sudo passwd rancher
.** Copy deployment file to VM
scp cloud-config.yml rancher@myip:~
Install Rancher OS to disk ( see documentation for more details ).
# With provisioning file
$ sudo ros install -c cloud-config.yml -d /dev/sda
# To allow local logins with rancher sudo user (this will allow local terminal login)
$ sudo ros install -d /dev/sda --append "rancher.password=mysecret"
INFO[0000] No install type specified...defaulting to generic
WARN[0000] Cloud-config not provided: you might need to provide cloud-config on boot with ssh_authorized_keys
Installing from rancher/os:v1.5.5
Continue [y/N]: y
INFO[0002] start !isoinstallerloaded
INFO[0002] trying to load /bootiso/rancheros/installer.tar.gz
69b8396f5d61: Loading layer [==================================================>] 11.89MB/11.89MB
cae31a9aae74: Loading layer [==================================================>] 1.645MB/1.645MB
78885fd6d98c: Loading layer [==================================================>] 1.536kB/1.536kB
51228f31b9ce: Loading layer [==================================================>] 2.56kB/2.56kB
d8162179e708: Loading layer [==================================================>] 2.56kB/2.56kB
3ee208751cd2: Loading layer [==================================================>] 3.072kB/3.072kB
Loaded image: rancher/os-installer:latest
INFO[0005] Loaded images from /bootiso/rancheros/installer.tar.gz
INFO[0005] starting installer container for rancher/os-installer:latest (new)
time="2020-04-14T19:07:05Z" level=warning msg="Cloud-config not provided: you might need to provide cloud-config on boot with ssh_authorized_keys"
Installing from rancher/os-installer:latest
mke2fs 1.45.2 (27-May-2019)
64-bit filesystem support is not enabled. The larger fields afforded by this feature enable full-strength checksumming. Pass -O 64bit to rectify.
Creating filesystem with 4193792 4k blocks and 4194304 inodes
Filesystem UUID: 409a9dd6-f91a-45f2-ab0a-772d52b1a390
Superblock backups stored on blocks:
32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632, 2654208,
4096000
Allocating group tables: done
Writing inode tables: done
Creating journal (16384 blocks): done
Writing superblocks and filesystem accounting information: done
Continue with reboot [y/N]: y
INFO[0046] Rebooting
INFO[0047] Setting reboot timeout to 60 (rancher.shutdown_timeout set to 60)
.............^[ ] reboot:info: Setting reboot timeout to 60 (rancher.shutdown_timeout set to 60)
.=.[ ] reboot:info: Stopping /docker : 42f28ebab847
............O............[ ] reboot:info: Stopping /open-vm-tools : 4fbf05cf6dca
..;..[ ] reboot:info: Stopping /ntp : 7e613153f723
..?..[ ] reboot:info: Stopping /network : 89a0f9f74c14
.......A.......[ ] reboot:info: Stopping /udev : a85c5a659768
..>..[ ] reboot:info: Stopping /syslog : 3391d49d2508
..=..[ ] reboot:info: Stopping /acpid : 4cb434297562
...C..[ ] reboot:info: Stopping /system-cron : dbace052a6d0
..J...[ ] reboot:info: Console Stopping [/console] : 7761e93fa8bb
Log in with your ssh key or use previously configured password.
Configure networking or use DHCP (enabled by default).
$ sudo ros config set rancher.network.interfaces.eth0.address 172.68.1.100/24
$ sudo ros config set rancher.network.interfaces.eth0.gateway 172.68.1.1
$ sudo ros config set rancher.network.interfaces.eth0.mtu 1500
$ sudo ros config set rancher.network.interfaces.eth0.dhcp false
$ sudo ros config set rancher.network.dns.nameservers "['192.168.4.1','192.168.2.1']"
- Validate new configuration.
$ sudo ros c validate -i /var/lib/rancher/conf/cloud-config.yml
# No errors means config is ok
- Change to console that supports state persistence.
$ sudo ros console switch alpine
Switching consoles will
1. destroy the current console container
2. log you out
3. restart Docker
Continue [y/N]: y
Pulling console (docker.io/rancher/os-alpineconsole:v1.5.5)...
v1.5.5: Pulling from rancher/os-alpineconsole
9d48c3bd43c5: Pull complete
cbc3261af7c6: Pull complete
5d1a52b7b5d3: Pull complete
Digest: sha256:39de2347295a5115800d6d9d6d747a8c91d2b06d805af927497754a043072188
Status: Downloaded newer image for rancher/os-alpineconsole:v1.5.5
Relogin to appliance
Reboot to apply network configuration with
sudo reboot
(if using static IP).Disable local password with
sudo ros config syslinux
and remove append part from configuration file (optional).Install appliance management and run installation and configuration wizard.
docker run -it \
--privileged \
--net=host \
-v /var/run/docker.sock:/var/run/docker.sock:rw \
-v /:/appliancehost \
notakey/cfg:latest install
Unable to find image 'notakey/cfg:latest' locally
latest: Pulling from notakey/cfg
c9b1b535fdd9: Pull complete
d1804e804407: Pull complete
Digest: sha256:8a9fb9f387c7f493a964cfad453924b6031935b737eab31ebaa8451c31df689f
Status: Downloaded newer image for notakey/cfg:latest
_______ __ __
\ \ _____/ |______ | | __ ____ ___.__.
/ | \ / _ \ __\__ \ | |/ // __ < | |
/ | ( <_> ) | / __ \| <\ ___/\___ |
\____|__ /\____/|__| (____ /__|_ \\___ > ____|
\/ \/ \/ \/\/
Welcome to the Notakey Authentication Appliance
installation wizard.
This wizard will install required utilities for appliance
management.
Do you wish to continue? (y/n): y
==> Detecting OS
Detected RancherOS version v1.5.5 from os image rancher/os:v1.5.5
==> Detecting console
==> Determining user
Assuming rancher
==> Creating directories
==> Installing management utilities
==> Installing default config
==> Installing ntk cli wrapper
==> Pulling the specified tag... skipping (--skip-pull detected)
==> Checking if tag exists locally... ok
==> Writing the wrapper... ok
==> Setting executable permission... ok
==> Installing coreutils...done
==> RancherOS configuration
==> Storing version information
==> Storing configuration
NOTE: Install can create user with limited privileges
that can be used for appliance daily management.
This user has only access to ntk prompt, docker and
few basic utilities.
Do you wish create limited user? (y/n): n
==> Configuring assets directory
Assets stored in /home/rancher/userlogos
Edit /opt/notakey/cfg/.env to adjust
==> Downloading docker images
==> The following package versions will be downloaded
downloading all latest packages
dashboard: dashboard:2.19.3
auth-proxy: authproxy:1.0.4
cli: cfg:5.0.0
consul: consul:1.5.2
traefik: traefik:1.7.21
vrrp: keepalived:2.0.20
htop: htop:latest
sso: sso-idp:2.4.17
==> Downloading from repo.notakey.com
downloading dashboard:2.19.3
downloading authproxy:1.0.4
downloading cfg:5.0.0
downloading consul:1.5.2
downloading traefik:1.7.21
downloading keepalived:2.0.20
downloading htop:latest
downloading sso-idp:2.4.17
==> Configuration update
update complete
Do you wish to run configuration wizard? (y/n): y
... wizard continues, see below for details
Appliance Configuration
Multiple user provisioning
If you wish to carry out administration tasks via CLI interface using different user contexts to avoid credential sharing you have two options:
all admins use rancher user context and log on using individual SSH public keys (this is the recommended RancherOS way)
create admin user for each individual user, each user then can setup individual SSH public key to log in.
Creating individual users
Run the following commands to add a new user with username admin1. Replace all instances of admin1 with username you wish to add.
sudo adduser -s /bin/bash admin1
sudo addgroup admin1 docker
Allow user sudo access to be able to add new admin users and make system service changes. Normally all daily operations can be
carried out with ntk
wrapper command. Add user to sudo only if user needs to invoke ros
manage RancherOS configuration. Remember to remove
rancher user password with sudo ros config syslinux
, if using password during setup.
# To allow root level access, add user to sudo group
sudo addgroup admin1 sudo
# Add user to sudoers file and change permissions
sudo sh -c 'echo "admin1 ALL=(ALL) ALL" > /etc/sudoers.d/admin1 && chmod 0440 /etc/sudoers.d/admin1'
User then can register his public key as follows.
cat ~/.ssh/id_rsa.pub | ssh admin1@nodeip 'mkdir -p .ssh && chmod 0700 .ssh && cat >> .ssh/authorized_keys'
This command assumes you SSH public key is in ~/.ssh/id_rsa.pub. Replace username admin1 and nodeip according to your setup.
To remove user run sudo deluser admin1
.
Assembling multi node cluster
Any licensing model allows cluster setup starting from two or more nodes. Nodes in cluster can be located in single virtual datacenter or can be deployed on multiple datacenters provided that there is a reliable communications link between nodes. If no L2 network cannot be provided between datacenters, advanced configuration is possible deploying a private routed network between nodes.
Configuration of first node
- Start with basic configuration wizard.
[ntkadmin@appserver-1/mydomain]$ ntk wizard
_______ __ __
\ \ _____/ |______ | | __ ____ ___.__.
/ | \ / _ \ __\__ \ | |/ // __ < | |
/ | ( <_> ) | / __ \| <\ ___/\___ |
\____|__ /\____/|__| (____ /__|_ \___ > ____|
\/ \/ \/ \/\/
Welcome to the Notakey Authentication Appliance
configuration wizard.
This wizard is meant to guide you through the initial
setup.
Advanced configuration is always available via the 'ntk'
command. Consult the user manual for more details.
==> Detecting network interfaces
==> Gathering information
NOTE: Datacenter / domain must be the same if assembling
local multi node cluster. Local cluster resides in
single LAN segment with unrestricted communication.
This value must contain only alphanumeric characters
and - with no other punctuation or spaces.
Enter the node hostname: appserver-1
NOTE: Datacenter / domain must be the same if assembling
local multi node cluster. Local cluster resides in
single LAN segment with unrestricted communication.
This value must contain only alphanumeric characters
and - with no other punctuation or spaces.
Enter the cluster datacenter name/local domain: mydomain
NOTE: You can specify below to join automatically an existing
cluster. Specifying y will start cluster locally and
attempt to join existing cluster by connecting to
specified IP address. Cluster service must be started on
remote host and must be in healthy state.
If unsure, say n, you will have a chance to do this later.
Are you joining existing cluster? (y/n) [n]: n
Enter the desired domain name (e.g. ntk.example.com): ntk.example.com
Detected interfaces: eth0
Enter interface you would like to configure: eth0
Use DHCP? (y/n): n
NOTE: This wizard does not support IPv6 IP settings
Consult the appliance user manual for more details on
using IPv6.
Enter the static IPv4 in CIDR notation (e.g. 10.0.1.2/24): 192.168.1.2/24
Enter the static IPv4 gateway (e.g. 10.0.1.1): 192.168.1.1
Enter the DNS server addresses (separate values with ,): 192.168.1.1
Use automated TLS (Lets Encrypt?) (y/n): y
Enter e-mail address to use for requesting TLS certificate: helpdesk@example.com
==> Please review the entered information
Network interface: eth0
Use DHCP: n
Static IPv4: 192.168.1.2/24
Static gateway: 192.168.1.1
DNS servers: 192.168.1.1
External domain: ntk.example.com
Use Lets Encrypt (automatic TLS): y
E-mail for Lets Encrypt: helpdesk@example.com
Are these values OK? (y/n): y
==> Applying configuration
acme... ok
network settings... ok
basic settings... ok
- Configure to run management on separate port, if required. Application accelerator (caching) has to be disabled if using management port.
[ntkadmin@appserver-1/mydomain]$ ntk cfg setc nas.mgmt_port 8443
[ntkadmin@appserver-1/mydomain]$ ntk cfg setc :nas.caching 0
- Apply settings by restarting appliance.
[ntkadmin@appserver-1/mydomain]$ ntk sys reboot
- Start application and datastore in single node mode.
[ntkadmin@appserver-1/mydomain]$ ntk as start
==> Validating Notakey Authentication Server configuration
Checking security settings
Checking virtual network configuration
required: traefik-net
network exists
Checking software images
required: notakey/dashboard:2.5.0
Checking host
==> Validating Notakey Authentication Server cluster configuration
Checking node name
Checking cluster size
Checking if this node is cluster root... yes (root)
Checking advertise IP
Checking data center name
Checking TLS
using Lets Encrypt certificates
IMPORTANT: make sure this instance is publicly accessible via ntk.example.com
=> Starting NAS data store
client address: 172.17.0.1
cluster advertise address: 192.168.1.2
cluster size (minimum): 1
is cluster root: 1
started bebf41523ad60c6a222ec5b34be7bb64fd94751a95392d188c9eaa38320fc732
=> Starting new NAS instance
net: traefik-net
image: notakey/dashboard:2.5.0
cluster.image: consul:0.7.5
host: naa.4.0.0.notakey.com
cluster backend: http://172.17.0.1:8500/main-dashboard
loglevel: info
env: production
started 4de38ddc53a013445ab2b2c3c4c33537af26e4e0e44f935c2bd1e000bec6df04
=> Starting the reverse proxy
ACME status: on
ACME e-mail: admin@mycompany.com
started 49fcfdada52f3d5495ae1e03b517c90578ed3dd768ee154cdf72357c898d5f7b
Subsequent node configuration
- As on first node, start with basic configuration wizard.
[ntkadmin@appserver-2/mydomain]$ ntk wizard
_______ __ __
\ \ _____/ |______ | | __ ____ ___.__.
/ | \ / _ \ __\__ \ | |/ // __ < | |
/ | ( <_> ) | / __ \| <\ ___/\___ |
\____|__ /\____/|__| (____ /__|_ \___ > ____|
\/ \/ \/ \/\/
Welcome to the Notakey Authentication Appliance
configuration wizard.
This wizard is meant to guide you through the initial
setup.
Advanced configuration is always available via the 'ntk'
command. Consult the user manual for more details.
==> Detecting network interfaces
==> Gathering information
NOTE: Node hostname must be unique within cluster and
must contain only alphanumeric characters and -
with no other punctuation or spaces.
Enter the node hostname []: appserver-2
NOTE: Datacenter / domain must be the same if assembling
local multi node cluster. Local cluster resides in
single LAN segment with unrestricted communication.
This value must contain only alphanumeric characters
and - with no other punctuation or spaces.
Enter the cluster datacenter name/local domain []: mydomain
NOTE: You can specify below to join automatically an existing
cluster. Specifying y will start cluster locally and
attempt to join existing cluster by connecting to
specified IP address. Cluster service must be started on
remote host and must be in healthy state.
If unsure, say n, you will have a chance to do this later.
Are you joining existing cluster? (y/n) [y]:
Please provide numeric IP adress for existing cluster member: 192.168.1.2
Detected interfaces: eth0
Enter interface you would like to configure: eth0
Use DHCP? (y/n): n
Enter the static IPv4 in CIDR notation (e.g. 10.0.1.2/24): 192.168.1.3/24
Enter the static IPv4 gateway (e.g. 10.0.1.1): 192.168.1.1
Enter the DNS server addresses (separate values with ,): 192.168.1.1
Use automated TLS (Lets Encrypt?) (y/n): y
Enter e-mail address to use for requesting TLS certificate: helpdesk@example.com
==> Please review the entered information
Network interface: eth0
Use DHCP: n
Static IPv4: 192.168.1.2/24
Static gateway: 192.168.1.1
DNS servers: 192.168.1.1
Joining existing cluster: 192.168.1.1
Are these values OK? (y/n): y
==> Applying configuration
acme... ok
network settings... ok
basic settings... ok
joining cluster... ok
- Management port should be set individually on each node, if required. Application accelerator (caching) has to be disabled if using management port.
[ntkadmin@appserver-2/mydomain]$ ntk cfg setc nas.mgmt_port 8443
[ntkadmin@appserver-2/mydomain]$ ntk cfg setc :nas.caching 0
- Apply settings by restarting appliance.
[ntkadmin@appserver-2/mydomain]$ ntk sys reboot
- Verify that the nodes have successfully synced.
[ntkadmin@appserver-2/mydomain]$ ntk cluster nodes
Node Status IP Address Data Center
appserver-1 alive 192.168.1.2 mydomain
appserver-2 alive 192.168.1.3 mydomain
- Start application on node.
[ntkadmin@appserver-2/mydomain]$ ntk as start
...
Further steps
To get the service fully up and running here is the list of steps required:
- VRRP configuration, if built in VRRP is used - please consult Virtual Router Redundancy Protocol (VRRP) section.
- Authentication server configuration - upload license, configure required services, onboarding requirements and user sources. Manual is available when logging on to administration dashboard.
- Notakey Authentication Proxy configuration - please consult Authentication Proxy section.
Appliance Upgrade
Subsections below describe various upgrade scenarios.
RancherOS upgrade
The following commands should be run as user rancher
. Limited user that can be created during installattion does not have the necessary privileges to perform OS upgrade/downgrade.
From: https://rancher.com/docs/os/v1.2/en/upgrading/
If RancherOS has released a new version and you want to learn how to upgrade your OS, it can be easily done with the ros os
command.
Since RancherOS is a kernel and initrd, the upgrade process is downloading a new kernel and initrd, and updating the boot loader to point to it. The old kernel and initrd are not removed. If there is a problem with your upgrade, you can select the old kernel from the Syslinux bootloader.
RancherOS version control
Use command below to get your current RancherOS version.
$ sudo ros os version
v1.5.6
RancherOS available releases
Use command below to get available RancherOS versions.
$ sudo ros os list
rancher/os:v1.5.6 local latest
rancher/os:v1.5.5 local available running
rancher/os:v1.5.4 remote available
rancher/os:v1.5.3 remote available
rancher/os:v1.5.2 remote available
RancherOS upgrade to the latest release
Use command below to upgrade your RancherOS to current version.
$ sudo ros os upgrade
Upgrading to rancher/os:v1.5.6
Continue [y/N]: y
...
...
...
Continue with reboot [y/N]: y
INFO[0037] Rebooting
RancherOS upgrade/downgrade to a specific version
If you’ve upgraded your RancherOS and something’s not working anymore, you can easily rollback your upgrade, just pick it up after -i option. If there is the new RancherOS version available but not listed yet, you can upgrade by manually picking it up after -i option.
$ sudo ros os upgrade -i rancher/os:v1.5.7
Upgrading to rancher/os:v1.5.7
Continue [y/N]: y
...
...
...
Continue with reboot [y/N]: y
INFO[0082] Rebooting
RancherOS user docker version change
User docker service is responsible for running all NtkAA services. Whenever you are running docker command directly from cli you are interfacing with this service. RancherOS vendor is recommending against docker version changes without OS upgrade so perform this operation at your own discretion and risk.
Use command sudo ros engine list -u
to discover available docker versions.
Use command sudo ros engine switch docker-19.03.11
to update to specific version.
4.2.x with single node.
Below are steps required to migrate previous generation appliance to current version.
Save backup with
ntk backup save /home/ntkadmin/migrate.backup
.Copy backup to your PC with:
scp ntkadmin@myip:/home/ntkadmin/migrate.backup ~/
Halt 4.2 version appliance with
ntk sys shutdown
.Install new appliance as described here.
Configure network identically as in previous version.
Run wizard during installation and configure node name and datacenter as specified in 4.2 appliance.
Copy backup file to new appliance
scp ~/migrate.backup rancher@myip:~/
.Load backup with
ntk backup load /home/rancher/migrate.backup
Start the required services manually (e.g.
ntk as start
).
4.2.x multinode
Steps below describe process of removal of single previous version node from cluster and joining a new version appliance in place.
Mixed version configuration, when some nodes in cluster are running older version, is generally not an issue, provided that docker images are regularly updated.
Update the node you plan to replace to latest versions with
ntk sys update
.Save backup with
ntk backup save /home/ntkadmin/migrate.backup
on the node you plan to replace.Copy backup to your PC with:
scp ntkadmin@myip:/home/ntkadmin/migrate.backup ~/
Halt 4.2 version appliance with
ntk sys shutdown
.Verify that cluster availability has not changed with
ntk cluster health
on other node in cluster.Install new appliance as described here.
Configure network identically as in previous version.
Run wizard during installation and configure different (unique) node name and identical datacenter.
Join node to cluster during wizard setup.
Copy backup file to new appliance
scp ~/migrate.backup rancher@myip:~/
.Load backup with
ntk backup load /home/rancher/migrate.backup --skip-net --skip-cluster
. This will restore only local non-network configuration.Start the required services manually (e.g.
ntk as start
).Verify that services have started successfully (
ntk <service> status
).
Configuration Management
Commands in section ntk cfg
provide unified interface for configuration management.
Command reference
ntk cfg getc
Usage:
ntk cfg getc [<cfg_key>]
ntk cfg getc [<cfg_key>] --batch
ntk cfg getc :
Description:
Retieve a configuration key. Configuration keys can be hierarchical, meaning that keys can contain subkeys.
Enter parent key to get overview on child configuration keys and values. Keys with colon (:) prefix are stored in cluster, so to
query them cluster has to be started and readable. Use two semicolons (::) to retrieve configuration values from
cluster and, if value is not present, fall back to locally stored value. Batch param –batch allows piping output
to file without color formatting. Using ntk cfg getc :
and specifying : as key param allows inspecting all cluster configuration keys.
ntk cfg setc
Usage:
ntk cfg setc <cfg_key> <value>
ntk cfg setc <cfg_key> <file> --load-file
ntk cfg setc <cfg_key> <value_struct> --json-input
Description:
Set configuration key to specified value. Configuration keys can be hierarchical, meaning that keys can contain subkeys.
Option –json-input can be used to populate child keys, parent value will be taken form specified
It is possible to add standard emojis to configuration values, for example, to generate authentication requests. Appliance stores all configuration strings using UTF-16 encoding internally, so codes for emoji should be formatted as UTF-16 character pairs and must be used in a valid combination. No character escaping is necessary.
Here is an example of configuring :sunglasses: emoji for NtkAP:
ntk cfg setc ap.message_description "Hello {0}, ready to do some off-site work \uD83D\uDE0E ?"
Partial vs Full Configuration Keys
You can retrieve values using their full key (e.g. node.name
) or you can
retrieve parts of the configuration key by specifying a partial key (e.g. just
node
for all the values underneath it).
# Example illustrating that configuration values can be retrieved
# using both full and partial keys.
# Full key
$ ntk cfg getc node.name
NotakeyNode1
# Full key
$ ntk cfg getc node.datacenter
DefaultDC
# Partial key will retrieve all keys underneath it
$ ntk cfg getc node
{
"name": "NotakeyNode1",
"datacenter": "DefaultDC"
}
# Store key in cluster configuration
$ ntk cfg setc :sso.host sso.example.com
sso.example.com
# Get key from cluster configuration
$ ntk cfg getc :sso.host
sso.example.com
# Store key locally and retrieve with fallback to local value
# If there is a cluster stored value, it will override local value
$ ntk cfg delc :sso.host
$ ntk cfg setc sso.host mysso.example.com
sso.example.com
$ ntk cfg getc ::sso.host
mysso.example.com
# Save boolean value in test.boolkey
$ ntk cfg setc test.boolkey true --json-input
true
# Check for saved value
$ ntk cfg getc test
{
"boolkey": true
}
# Save integer value in test.intkey
$ ntk cfg setc test.intkey 98467 --json-input
98467
# Check for saved value
$ ntk cfg getc test
{
"boolkey": true,
"intkey": 98467
}
Setting Nested Values (--json-input)
An important parameter to know is --json-input
, which is sometimes necessary
for setting values. It directs the configuration engine to interpret the configuration
value as JSON values. This allows setting multiple nested values at once.
By default, values will be interpreted as a string or a number. This is usually fine. However, when using a partial configuration key, it may override values nested below. For example:
# Example of how node.name and node.datacenter
# values can be overridden by accident.
$ ntk cfg setc node { "name": "NodeName", "datacenter": "NodeDC" }
{ name: NodeName, datacenter: NodeDC }
# Without --json-input, the value was
# interpreted as one long string
$ ntk cfg getc node
{ name: NodeName, datacenter: NodeDC }
# The nested value node.name does not exist
# anymore
$ ntk cfg getc node.name
Alternatively, if we use --json-input
, then the values are set hierarchically:
# Note the use of single quotes around
# the value
$ ntk cfg setc node '{ "name": "NodeName", "datacenter": "NodeDC" }' --json-input
{
"name": "NodeName",
"datacenter": "NodeDC"
}
# Multi-line output hints that the value
# is not a simple string anymore
$ ntk cfg getc node
{
"name": "NodeName",
"datacenter": "NodeDC"
}
$ ntk cfg getc node.name
NodeName
Loading Values From a File (--load-file)
You can load values from a file by passing the --load-file
argument
to ntk cfg setc...
. The value argument will be interpreted as a path to file,
and the contents of this file will be used as the real configuration value.
ntk cfg setc <key> <path_to_file> --load-file
$ echo "Hello World" > /tmp/value_file
$ ntk cfg setc test_value /tmp/value_file --load-file
Hello World
$ ntk cfg getc test_value
Hello World
The parameter is useful when setting these configuration keys, which are cumbersome to set directly:
Key | Description |
---|---|
:ssl.pem | HTTPS certificate in PEM format |
:ssl.key | Certificate key in PEM format |
ntk cfg migrate
Usage:
ntk cfg migrate <cfg_key_src> <cfg_key_dst>
Description:
Retrieve configuration key and store into destination key. Use :keyname for global cluster backed configuration keys and keyname without any prefix for local persistent configuration. If source key is empty, destination value will be effectively deleted.
Task Scheduler
Commands in ntk cron
enable or disable running of maintenance jobs in subscribed services. This must be enabled on at least one of nodes in cluster and can be safely enabled on all nodes.
Services like NAS, SSO subscribe to incoming events in cron and run their maintenance jobs as needed in daily / hourly / weekly and mothly schedule.
Command reference
ntk cron enable
Usage:
ntk cron enable
Description:
Enables cron task processing globally on node. Full node restart required.
ntk cron disable
Usage:
ntk cron disable
Description:
Disables global cron task processing on node.
ntk cron run
Usage:
ntk cron run
Description:
Allows you to debug tasks processing. Output will show tasks that will be run and their output.
Reverse Proxy
Commands in ntk rp
section manage Reverse Proxy configuration and status. Reverse Proxy handles all incoming HTTP(S) requests and routes them to corresponding providers.
Configuration settings
The following configuration settings can be configured:
Key | Description | Default |
---|---|---|
:ssl.acme.status | On/off flag. When on - TLS certificates will be created and auto-renewed, using Let’s Encrypt. When off, TLS certificates must be set via the ssl.pem and ssl.key configuration values. |
off |
:ssl.acme.email | When using Let’s Encrypt TLS certificates, this will be sent as the contact address. | - |
:ssl.acme.mode | When using Let’s Encrypt TLS certificates, you can use either HTTP-01 (config value http) or TLS-01 methods (config value tls). Default is http. | http |
:ssl.pem | Default TLS certificate (in PEM format). This is used only when ssl.acme.status is not on . You can set this value directly. It should begin with -----BEGIN CERTIFICATE----- and have newlines escaped as \n . This value usually requires certification chain, so there can be multiple certificates, starting from intermediate. |
- |
:ssl.key | Default TLS certificate key (in PEM format). This is used only when ssl.acme.status is not on . You can set this value directly. It should begin with -----BEGIN PRIVATE KEY----- and have newlines escaped as \n . |
- |
:ssl.acme.mgmt_enable | On/off flag. Enable management interface TLS certificate acquired from ACME. If ACME is used and this is not enabled, management will use self signed certificate. | off |
:ssl.acme.ca | ACME CA URL. Default value is production URL (https://acme-v02.api.letsencrypt.org/directory), use staging https://acme-staging-v02.api.letsencrypt.org/directory for tests. | |
ssl.acme.debug | Set to "true” to enable ACME certificate provisiosning debuging. This is a local setting. | - |
:rp.custom_params | Key value pairs of custom configuration options, e.g. key1=value,key2=value . |
- |
:rp.extra_domains | Additional domains for local custom services. Comma seperated values e.g. “myapp.example.com,other.example.com”. HTTP TLS certificate will be requested from ACME for all domains listed here. | - |
Change settings using ntk cfg setc
command.
Command reference
ntk rp restart
Usage:
ntk rp restart
Description:
Stops and start Reverse Proxy. Use this command to apply configuration changes.
ntk rp start
Usage:
ntk rp start [--force]
Description:
Starts Reverse Proxy. Use –force option to skip TLS certificate validation matching configured DNS hostname e.g. useful during initial setup.
ntk rp stop
Usage:
ntk rp stop
Description:
Stops Reverse Proxy. Service will be stopped and will not be started until ntk as start
is called. The
stopped state persists across appliance restarts.
ntk rp resettls
Usage:
ntk rp resettls
Description:
Resets all ACME TLS settings and starts certificate acquisition and validation from scratch.
ntk rp status
Usage:
ntk rp status [--short]
Description:
Shows status of all Reverse Proxy components. Use –short to display only status (useful for scripting).
ntk rp validate
Usage:
ntk rp validate
Description:
Validates Reverse Proxy configuration, checks TLS certificate validity, cluster and networking configuration.
Authentication Server
Commands in ntk as
section govern Notakey Authentication Server (NtkAS) state and configuration. Notakey Authentication Server consists of three distinct
parts - device API used for interacting with mobile clients, management API used for managing services and sending authentication requests and finally
management browser based user interface. When installing appliance all components are activated and available on https port of appliance
Configuration settings
The following configuration settings must be present to start this service.
Key | Description | Default |
---|---|---|
:nas.host | External FQDN for authentication API service. | - |
nas.mgmt_port | Set port number, used for management, see below for details. | 443 |
nas.auto_usersync | Set to 1 to enable automatic user sync for all services from external user repository sources (e.g. LDAP and ActiveDirectory). | 0 |
:nas.update_notification | Set to 0 to disable automatic update notifications for Notakey Authentication Server for all users in Administration Service | 1 |
:nas.caching | Set to 0, to disable caching (caching will not work correctly in load balanced environments). | 1 |
:nas.cache_ttl | Set lifetime of cached entries in seconds. | 3600 |
:nas.redis_url | URL for REDIS application accelerator distributed cache. Add multiple (sharded) servers divided by , e.g. redis://redis1:6379/0,redis://redis2:6379/0 |
- |
:messenger_url | Set URL used for push notification and SMS sending. Set to off to disable push notifications. | https://messenger.notakey.com |
nas.healthcheck | Set to “off” to disable internal service consistency check. Use only if experiencing issues with service becoming unhealthy due unhandled spikes in load. | on |
:nas.healthcheck_interval | Interval in seconds between health checks | 10 |
:nas.healthcheck_timeout | Timeout in seconds for health probe | 10 |
:nas.auth.ttl | API access token lifetime in seconds. When token expires, new has to be requested. | 7200 |
:nas.ks.pubkey_ttl | Lifetime in seconds for keys used for encryption. | 3600 |
:nas.log.retention.auth_log | Days to store authentication log in cluster datastore. Use syslog audit endpoint for long term storage. | 30 |
:nas.log.retention.onboarding_log | Days to store onboarding log in cluster datastore. Use syslog audit endpoint for long term storage. | 90 |
:nas.log.retention.auth_request | Days to store processed, signed authentication requests. | 30 |
:nas.phone_country_code | Phone number prefix to attempt to combine with user source provided number to get a valid number in international format. If phone numbers already are in international format, this setting is ignored. | 371 |
:nas.syslog | IP address or hostname to send technical logs. For high volume nodes, IP address is recommended. | - |
:nas.syslog_level | Loglevel number to enable logging for. All events in this log level or above will be sent to syslog. The available log levels are: debug, info, warn, error, fatal, and unknown, corresponding to the log level numbers from 0 up to 5, respectively. | 3 |
:auth_domain | Domain of authentication server (as configured using DNS SSD). If none is specified, fallback value is used based on main host (ntk cfg get :nas.host). | http://<host> |
:ssl.nas.pem | TLS certificate (in PEM format). This is used only when ssl.acme.status is not on . You can set this value directly. It should begin with -----BEGIN CERTIFICATE----- and have newlines escaped as \n . This value usually requires certification chain, so there can be multiple certificates, starting from intermediate. |
- |
:ssl.nas.key | TLS certificate key (in PEM format). This is used only when ssl.acme.status is not on . You can set this value directly. It should begin with -----BEGIN PRIVATE KEY----- and have newlines escaped as \n . |
- |
:nas.custom_params | Key value pairs of custom configuration options, e.g. key1=value,key2=value . |
- |
:nas.secret_key_base | Management session cookie encryption key (generated automatically on first run). | - |
:nas.external_proxy | Set to “on” to expose NtkAS directly to external proxy / load balancer. Service will run on tcp/5000. | off |
:nas.expose_mgmt_api | Set to “off” to disable access to management API over the regular HTTPS port. | on |
:nas.dual_instance_mode | Set to “on” to enable legacy dual container mode, where management runs on seperate container. If :nas.external_proxy is enabled, service will run on tcp/6000. |
off |
Change settings using ntk cfg setc <Key> <value>
command.
Running management dashboard on non-default port
It is possible to configure management dashboard on port other than default 443, that is used for device API. To set this to
different port run e.g. ntk cfg setc nas.mgmt_port 8443
, so that management data is separated from payload data. By default all management
APIs are also published on public HTTPS endpoint. Configure ntk setc :nas.expose_mgmt_api off
followed by NtkAS restart to disable public management API access.
Remote management API consumers will have to use specified management port for management operations (e.g. creating users, sending authentication requests).
HTTPS certificate for management is set with ntk cfg setc :ssl.pem <cert>
and ntk cfg setc :ssl.key <key>
. ACME certificate configuration on non default port is not possible.
Command reference
ntk as restart
Usage:
ntk as restart
Description:
Stops and start Notakey Authentication Server. Use this command to apply configuration changes.
ntk as start
Usage:
ntk as start [--force]
Description:
Starts Notakey Authentication Server. Use –force option to skip TLS certificate validation matching configured DNS hostname e.g. useful during initial setup.
ntk as stop
Usage:
ntk as stop
Description:
Stops Notakey Authentication Server. Service will be stopped and will not be started until ntk as start
is called. The
stopped state persists across appliance restarts.
ntk as status [–short]
Usage:
ntk as status
Description:
Shows status of all Notakey Authentication Server components.
ntk as validate
Usage:
ntk as validate
Description:
Validates Notakey Authentication Server configuration, checks TLS certificate validity, cluster and networking configuration.
ntk as auth
Usage:
ntk as auth add <client name> [<client ID> <client secret>]
ntk as auth del <client ID>
ntk as auth scopes <client ID> "<scope1 scope2 scope3>"
ntk as auth ls
Description:
Manages M2M client credentials, that can be used to manage resources using APIs provided by Notakey Authentication Server.
See Notakey API integration documentation for details how to establish authentication session using these credentials.
Use ntk as auth add
to add a new API consumer cient. Client ID and client secret params are optional and will be generated automatically if not provided.
To define access permissions use ntk as auth scopes
command. Scopes are defined per client and have to be requested for every session explicitly.
Refer to Notakey API integration documentation for details.
ntk as report
Usage:
ntk as report services
ntk as report users <service ID>
ntk as report devices <service ID>
Description:
Exports tab seperated report of configured NAS services, users and devices. Use ntk as report services
to first discover the “service ID” parameter for users and devices
commands.
ntk as bootstrap
Usage:
ntk as bootstrap
Description:
Provision authentication server with initial root application, admin team, single admin user and license. Use this wizard if no direct web manangement UI is accesible during install or setting up the service to be run through external load balancer.
[root@node/dc1]$ ntk as bootstrap
NOTE: This command provisions Notakey Authentication Server. All further
configuration must be done using web UI, which is running
either on standard https port or one specified in mgmt_port
config entry.
Please enter admin username: admin
Please enter admin password:
Retype new password:
Do you wish to disable 2FA for root service? (y/n): y
Provide full path to P12 licence file: /home/ntkadmin/license.p12
Proceed with recovery? (y/n): y
...
ntk as recoveradm
Usage:
ntk as recoveradm
Description:
Recovers admin user access to root administration service and enables onboarding for administration service. Any active existing administrators will receive notification of this change.
[root@node/dc1]$ ntk as recoveradm
NOTE: This command recovers admin user access to root administration
service by adding new admin user, enabling onboarding for
administration service and optionally changing 2FA requirement
for administrator access. Any active existing administrators will
receive notification about this change.
Please enter admin username: admin
Please enter admin password:
Retype new password:
Do you wish to disable 2FA for root service? (y/n): y
Proceed with recovery? (y/n): y
...
Single Sign-On Server
Notakey Single Sign-On Server (NtkSSO) is a SAML, OAuth 2.0 and various other authentication protocol capable solution for web and mobile authentication workflows based on SimpleSAMLphp (SSP) project.
SSP Commands in ntk sso
section govern NtkSSO state and configuration.
Configuration settings
Key | Type | Description |
---|---|---|
:sso.host | string | FQDN by which will be used by users to access SSO service (e.g. sso.example.com). If ssl.acme.status is on , then this should be a publicly accessible address. |
:sso.base.“admin.protectindexpage” | bool | Do not allow direct access to admin startpage without authentication, true / false value. |
:sso.base.“admin.protectmetadata” | bool | Do not allow direct anonymous access to metadata used by service providers, true / false value. |
:sso.base.“auth.adminpassword” | string | Sets local admin password, allows acess to metadata and diagnostics. |
:sso.base.errorreporting | bool | Enables user error reporting form when problems occur, true / false value. |
:sso.base.technicalcontact_name | string | Contact name for error reporting form. |
:sso.base.technicalcontact_email | string | Email where to send error reports from error reporting form. |
:sso.base.debug | bool | Enable debugging of authentication processes. |
:sso.base.showerrors | bool | Enable or disable error display to end user (in form of backtrace). |
:sso.base.“logging.level” | integer | Sets log level, integer value from 0 to 7, 7 being most verbose, used with debug. |
:sso.base.“theme.use” | string | Sets custom theme, string value in format “module:theme”. |
:sso.base.“session.duration” | integer | Maximum duration of authenticated session in seconds, default 8 hours. |
:sso.base.“session.cookie.name” | string | Option to override the default settings for the session cookie name. Default: ‘NTK-SSO’ |
:sso.base.“session.cookie.lifetime” | integer | Expiration time for the session cookie, in seconds. Defaults to 0, which means that the cookie expires when the browser is closed. Default: 0 |
:sso.base.“session.cookie.path” | string | Limit the path of the cookies. Can be used to limit the path of the cookies to a specific subdirectory. Default: ‘/’ |
:sso.base.“session.cookie.domain” | string | Cookie domain. Can be used to make the session cookie available to several domains. Default: nil |
:sso.base.“session.cookie.secure” | bool | Set the secure flag in the cookie. Set this to TRUE if the user only accesses your service through https. If the user can access the service through both http and https, this must be set to FALSE. Default: FALSE |
:sso.base.“language.default” | string | Sets default ISO 639-1 language code for UI language. |
:sso.base.“trusted.url.domains” | array | Array of domains that are allowed when generating links or redirections to URLs. simpleSAMLphp will use this option to determine whether toto consider a given URL valid or not, but you should always validate URLs obtained from the input on your own (i.e. ReturnTo or RelayState parameters obtained from the $_REQUEST array). Set to NULL to disable checking of URLs. simpleSAMLphp will automatically add your own domain (either by checking it dynamically, or by using the domain defined in the ‘baseurlpath’ directive, the latter having precedence) to the list of trusted domains, in case this option is NOT set to NULL. In that case, you are explicitly telling simpleSAMLphp to verify URLs. Set to an empty array to disallow ALL redirections or links pointing to an external URL other than your own domain. Example: [‘sp.example.com’, ‘app.example.com’]. Default: NULL |
:sso.base.“authproc.idp” | hash | Objects containing applied filters during authentication process, see below for details. |
:sso.auth | hash | Contains objects that describe access to user repositories. At least one user repository must be configured. Please see below for details. |
:sso.“saml-sp” | hash | Objects describing all service provider (SP) metadata - these are web applications that consume the identity assertions. Please see below for details. |
:sso.“saml-idp” | hash | Objects describing names for this IdP. Only one IdP must be defined. Please see below for details. |
:sso.certificates | hash | Objects describing all certificates used by SSO e.g. used for SAML assertion signing. Please see below for details. |
:sso.modules | array | List of enabled modules. Only required modules should be enabled as this affects performance. Please see below for details. |
sso.healthcheck | string | Set to off to disable internal healthcheck. To be used with caution in cases when using external load balancer. |
:sso.healthcheck_interval | Interval in seconds between health checks | 10 |
:sso.healthcheck_timeout | Timeout in seconds for health probe | 10 |
:sso.cron | hash | Automatic task configuration. |
:sso.cron.mode | string | Set to “http” to enable legacy cron mode, useful for sso versions prior 3.0.0. |
:ssl.sso.pem | string | TLS certificate (in PEM format). This is used only when ssl.acme.status is not on . You can set this value directly. It should begin with -----BEGIN CERTIFICATE----- and have newlines escaped as \n . This value usually requires certification chain, so there can be multiple certificates, starting from intermediate. |
:ssl.sso.key | string | TLS certificate key (in PEM format). This is used only when ssl.acme.status is not on . You can set this value directly. It should begin with -----BEGIN PRIVATE KEY----- and have newlines escaped as \n . |
:sso.custom_params | string | Key value pairs of custom configuration options (cluster value with fallback to local node value), e.g. key1=value,key2=value . |
:sso.external_proxy | string | Set to “on” to expose NtkSSO directly to external proxy / load balancer. Application will be reachable on tcp/7000. |
Most configuration entries support storage in cluster, so prefix entries with : to share configuration between nodes. Exception is sso.image, that defines version of SSO image and if installing custom version, you will need to configure this on each node individually.
Change settings using ntk cfg setc
command. Use ntk cfg setc :sso.<key> <true|false> --json-input
format for boolean, integer and json structure values (hash and array) and ntk cfg setc :sso.<key> "<value>"
string values. See examples below on
implementation details for structure value settings.
Hashed passwords
For production use, any local passwords should be entered as hashes improving security. To generate password hashes use the following command in NtkAA cli terminal:
docker exec -it sso /var/simplesamlphp/bin/pwgen.php
Enter password: mystrongpassword
$2y$10$kQ/zxPQyfkb/bDPoKJ2es.7CRf5FGb2N5mwethSaKM.u5SWz62bFq
Administration access
To access various administration helper tools and inspect configured service providers you will need to access management interface.
This can be done by accessing url https://
To configure access to management interface you have to define :sso.auth.admin
authentication source.
Authentication with static password:
# define auth source
ntk cfg set :sso.auth.admin '{"module": "core:AdminPassword"}' --json-input
# generate password hash
docker exec -it sso /var/simplesamlphp/bin/pwgen.php
# enter generated hash into configuration
ntk cfg set :sso.base.\"auth.adminpassword\" '$2y$10$kQ/zxPQyfkb/bDPoKJ2es.7CRf5FGb2N5mwethSaKM.u5SWz62bFq'
# restart sso service
ntk sso restart
Authenticate with Notakey Authentication server module:
# define auth source
ntk cfg set :sso.auth.admin '{
"module": "notakey:Process",
"endpoints": [
{
"name": "SSO Admin Login",
"url": "https://ntk-api.host/",
"service_id": "bc...00",
"client_id": "...",
"client_secret": "c7...53",
"profile_id": "b3...29", // optional
"service_logo": "https://ntk-api.host/_cimg/b5..04?size=300x150"
}
]
}' --json-input
# restart sso service
ntk sso restart
See SSP documentation on authentication source configuration.
Userlogos directory
Any static file assets like images, certificates and custom css files can be stored in userlogos directory. Usually it is located in /home/ntkadmin/userlogos/, but can be changed by editing /opt/notakey/cfg/.env file. Contents of this folder will be synced automatically over to other nodes, potentially overwriting older contents. See “ntk sso sync” for more details.
Basic service configuration
Basic settings
$ ntk cfg setc :sso.host sso.example.com
sso.example.com
$ ntk cfg setc :sso.base.technicalcontact_name "Example.com support"
Example.com support
$ ntk cfg setc :sso.base.technicalcontact_email support@example.com
support@example.com
$ ntk cfg setc ':sso.base."auth.adminpassword"' "myadminpass"
myadminpass
Configuring user repository access
NAS user repository
$ ntk cfg setc :sso.auth.ntkauth '{ "module": "notakey:Process",
"endpoints": [
{
"name": "SSO service",
"url": "https://demo.notakey.com/",
"service_id": "bc...00",
"client_id": "a0...162",
"client_secret": "secret",
"profile_id": "b3d...529",
"service_logo": "https://demo.notakey.com/_cimg/b5d3d0f7-e430-4a30-bd34-cea6bab9c904?size=300x150"
}
]}' --json-input
Description:
Configure a user repository from Notakey Authentication Server.
Description of parameters:
Key | Type | Mandatory | Description |
---|---|---|---|
endpoints.?.name | string | yes | Name can be any string and will be shown on authentication request in title field. |
endpoints.?.url | string | yes | Service ID is the value issued after service creation and shown in NtkAA admin dashboard. |
endpoints.?.service_id | string | yes | Service ID of NtkAS, can be found on service page as “Access ID”. |
endpoints.?.client_id | string | yes | NtkAS management API OAuth client ID. |
endpoints.?.client_secret | string | yes | NtkAS management API OAuth client secret. |
endpoints.?.profile_id | string | no | NtkAS authentication profile ID to use for this endpoint. |
endpoints.?.service_logo | string | no | Relative or full URL to logo of this service. Use ~/userlogos directory to store local images. |
endpoints.?.stepup-source | string | no | Authentication source (from :sso.auth.xyz, where xyz is the authentication source name) to use for stepup authentication after 2FA verification. Usually some form of password verification. |
endpoints.?.stepup-duration | string | no | ISO 8601 formatted duration of how long the stepup authentication is valid, for example P30D for 30 days. |
user_id.attr | string | no | User attribute in filter mode to use for authentication. This attribute value will be used as NtkAS service username that will receive authentication request. |
LDAP based user repository
$ ntk cfg setc :sso.auth.myauth '{ "module": "ldap:LDAP",
"hostname": "ldaps://192.168.1.2",
"enable_tls": true,
"attributes": [
"sAMAccountName",
"cn",
"sn",
"mail",
"givenName",
"telephoneNumber",
"objectGUID"
],
"timeout": 10,
"referrals": false,
"dnpattern": "sAMAccountName=%username%,OU=MyOU,DC=example,DC=com"}' --json-input
Description:
Configure a user repository from LDAP, in this example from Microsoft Active Directory. Username matching is done by sAMAccountName attribute, but this could be configured as any other unique attribute. “search.attributes” defines which attributes are used for user ID matching.
SSO HTTP Certificate Configuration
To configure static TLS certificates for SSO HTTPS follow these steps:
Use scp or similar utility to transfer certificate and key files to appliance.
Use
ntk cfg setc :ssl.sso.pem
andntk cfg setc :ssl.sso.key
to load data from files into appliance configuration as shown below:
ntk cfg setc :ssl.sso.pem /home/ntkadmin/ssossl.pem --load-file 1> /dev/null
ntk cfg setc :ssl.sso.key /home/ntkadmin/ssossl.key --load-file 1> /dev/null
- Ensure that ACME certificates are disabled (
ntk cfg getc :ssl.acme.status
returns “off”).
Configuring Identity Provider
Typical ID provider configuration
$ ntk cfg setc :sso.\"saml-idp\".\"sso.example.com\" '{
"host": "__DEFAULT__",
"privatekey": "server.key",
"certificate": "server.pem",
"auth": "myauth"
}' --json-input
Description:
Configures an IdP with defined certificate and selected user repository. If multi-tenant mode is used, multiple IdPs can be configured and selection will be done
by HTTP host header matching to “host” configuration parameter. Fallback value is DEFAULT as in configuration above. The key and certificate has to be set with
command ntk cfg setc sso.certificates.server.key
and ntk cfg setc sso.certificates.server.pem
respectively. Name for certificate can be any valid filename.
Configuring service providers
Generic Google Gapps configuration
$ ntk cfg setc :sso.\"saml-sp\".\"google.com\" '{
"AssertionConsumerService": "https://www.google.com/a/example.com/acs",
"NameIDFormat": "urn:oasis:names:tc:SAML:1.1:nameid-format:emailAddress",
"simplesaml.attributes": false,
"startpage.logo": "gapps.png",
"startpage.link": "http://mail.google.com/a/example.com",
"name": "Google Gapps"
}' --json-input
Description:
Google Gapps for business customers can be configured to authenticate to remote SAML IdP. Here is the server side configuration for this setup.
Use ntk cfg delc sso.\"saml-sp\".\"google.com\"
to delete configured SP.
Amazon AWS console configuration
$ ntk cfg setc :sso.\"saml-sp\".\"urn:amazon:webservices\" '{
"entityid": "urn:amazon:webservices",
"OrganizationName": { "en": "Amazon Web Services, Inc." },
"name": { "en": "AWS Management Console" },
"startpage.logo": "aws-logo.png",
"startpage.link": "https://sso.example.com/sso/saml2/idp/SSOService.php?spentityid=urn:amazon:webservices",
"OrganizationDisplayName": { "en": "AWS" },
"url": { "en": "https://aws.amazon.com" },
"OrganizationURL": { "en": "https://aws.amazon.com" },
"metadata-set": "saml20-sp-remote",
"AssertionConsumerService": [ {
"Binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST",
"Location": "https://signin.aws.amazon.com/saml",
"index": 1,
"isDefault": true
} ],
"SingleLogoutService": "",
"NameIDFormat": "urn:oasis:names:tc:SAML:2.0:nameid-format:transient",
"attributes.required": [ "https://aws.amazon.com/SAML/Attributes/Role", "https://aws.amazon.com/SAML/Attributes/RoleSessionName" ],
"keys": [ {
"encryption": false,
"signing": true,
"type": "X509Certificate",
"X509Certificate": "MIIDbTCCAlWgAwIBAgIEKJJQLzANBgkqhkiG9w0BAQsFADBnMR8wHQYDVQQDExZ1\ncm46YW1hem9uOndlYnNlcnZpY2VzMSIwIAYDVQQKExlBbWF6b24gV2ViIFNlcnZp\nY2VzLCBJbmMuMRMwEQYDVQQIEwpXYXNoaW5ndG9uMQswCQYDVQQGEwJVUzAeFw0x\nNzA2MDYwMDAwMDBaFw0xODA2MDYwMDAwMDBaMGcxHzAdBgNVBAMTFnVybjphbWF6\nb246d2Vic2VydmljZXMxIjAgBgNVBAoTGUFtYXpvbiBXZWIgU2VydmljZXMsIElu\nYy4xEzARBgNVBAgTCldhc2hpbmd0b24xCzAJBgNVBAYTAlVTMIIBIjANBgkqhkiG\n9w0BAQEFAAOCAQ8AMIIBCgKCAQEAnmVupB5TWNR4Jfr8xU0aTXkGZrIptctNUuUU\nfA5wt7lEAgoH7JeIsw7I3/QYwfA0BkGcUxLiW4sj7vqtdXwLK5yHHOg3NYsN2wCU\noSnsw/msH7nyeyUXXRr9f+dyJRRViBnlaJ3RHNk6UZEbS/DJG/W4OMlthdzIXJbA\n0G0xQid96lUZOgeLuz1olWiAIbJRPqeBRcVlY8MvHzVE82ufvSgkGnMZAt6B4Q6T\n9PsIDdhRDbTBTtUAmaKF3NXl/IPq5eybM6Pw3gcVlsrtgq0LHsdhgRdMz6UeBaA9\nTn/dgmeWM+UKEXkHbtnsq6q9X+8Nq/HwGylI/p3cPZ8WuOjXeQIDAQABoyEwHzAd\nBgNVHQ4EFgQU7F4XurLqS7Bsxj8NKcA9BvHkEE0wDQYJKoZIhvcNAQELBQADggEB\nAJItHBnpD62VaOTSPqIzVabkpXWkmnclnkENAdndjvMcLJ3qrzqKb3xMQiEGCVzF\ni9ZXU3ucMAwufIoBaSjcuFNYT1Wuqn+Jed2xH5rUOAjeDndB7IjziU5kyW+6Zw9C\nFmSN8t0FFW2f3IALgWfrrJ6VWPzen9SJrNPmVralau66pDaZbGgpBkVi2rSfJiyM\nleau9hDNBhV+oPuvwDlbceXOdmJKUTkjyID9r+CmtHq7NL6omIH6xQw8DNFfa2vY\n17w2WzOcd50R6UUfF3PwGgEyjdiigWYOdaSpe0SWFkFLjtcEGUfFzeh03sLwOyLw\nWcxennNbKTXbm1Qu2zVgsks="
} ],
"saml20.sign.assertion": true,
"authproc": {
"91": {
"class": "core:AttributeAdd",
"https://aws.amazon.com/SAML/Attributes/Role": [ "arn:aws:iam::540274555001:role/SamlRoleAdmin,arn:aws:iam::540274555001:saml-provider/NTKSSO" ]
},
"92": {
"class": "core:PHP",
"code": "$attributes[\"https://aws.amazon.com/SAML/Attributes/RoleSessionName\"][0] = $attributes[\"uid\"][0];"
}
}
}' --json-input
Description:
Example above shows a more complex scenario where SP requires specific attributes present in assertion. The role SamlRoleAdmin therefore has to be created in AWS IAM module configuration. RoleSessionName in this case is simply copied from user ID value, a more elaborate setup would be required in production environment.
Configuring filters
Setup additional 2FA
$ ntk cfg setc :sso.base.\"authproc.idp\" '{
"90": {
"class": "notakey:Filter",
"user_id.attr": "sAMAccountName",
"endpoints": [
{
"name": "Mobile ID validation",
"url": "https://demo.notakey.com/",
"service_id": "bcd05d09-40cb-4965-8d94-da1825b57f00",
"client_id": "zyx",
"client_secret": "xyz",
"service_logo": "https://demo.notakey.com/_cimg/b5d3d0f7-e430-4a30-bd34-cea6bab9c904?size=300x150"
}
]
}
}' --json-input
Description:
Once you have configured a user repository as authentication backend in Identity Provider configuration, it is possible to add additional checks to authentication flow. Example above shows adding a NtkAA based backend for ID validation. Key “user_id.attr” is used to discover matching user in NtkAA, in this case sAMAccountName from Active Directory. It is possible to use additional modules and filters for languange provisioning with service providers, attribute mapping and alteration, etc.
Managing certificates
Adding a certificate
$ ntk cfg setc :sso.certificates.server '{
"description": "Demo certificate",
"pem": "-----BEGIN CERTIFICATE-----\nMIIDdDCCAlwCCQC30UW2PPlkQjANBgkqhkiG9w0BAQsFADB8MQswCQYDVQQGEwJM\nVjETMBEGA1UECBMKU29tZS1TdGF0ZTENMAsGA1UEBxMEUmlnYTEQMA4GA1UEChMH\nTm90YWtleTEWMBQGA1UEAxMNbm90YWtleS5sb2NhbDEfMB0GCSqGSIb3DQEJARYQ\naW5mb0Bub3Rha2V5LmNvbTAeFw0xNzAzMjExMjM4MzZaFw0yNzAzMTkxMjM4MzZa\nMHwxCzAJBgNVBAYTAkxWMRMwEQYDVQQIEwpTb21lLVN0YXRlMQ0wCwYDVQQHEwRS\naWdhMRAwDgYDVQQKEwdOb3Rha2V5MRYwFAYDVQQDEw1ub3Rha2V5LmxvY2FsMR8w\nHQYJKoZIhvcNAQkBFhBpbmZvQG5vdGFrZXkuY29tMIIBIjANBgkqhkiG9w0BAQEF\nAAOCAQ8AMIIBCgKCAQEA5L/rjXT0dEBmN3BSnsmLCwF9goOZWb5RNvvc+dnUQnaG\n5nJ8xMPSpEAm4y2ZlC8Uhm56H+7A2cUf/ogkUcKkyQHyXQrjjPzec5h3+734Mmit\nvHAAzC/j9smDOvdde9evMUDH9Sn7oJtpz+MKckTFvsBoLGuBPiuLn8P9HCmLKlMN\njR2hfJab+DZG+eysAGkoiGS9zOxMVF9Xsvk/EHxm5p3ZPF2oOEr8cimgQO/ew29+\nfF4etwskre/AA8KCw/6Bg2XwZ3HmFPJ9nd/zJSzqm5U96WFEth4ABVqZE4MS4rRo\n2K+eQy2rXzDnmiL3cGiLuL9aal+9JCFRxKE1RSIdEQIDAQABMA0GCSqGSIb3DQEB\nCwUAA4IBAQC7T9Yqu0+ONT/MP8ZNmbUs5B6+/S0yIreLC1pMmJymxMDIFQCnxeRf\naxDtDfEJVzkiMAl2Q3q/HApCDh/l+hpbn4oKoDXTQvWiZpOWhSR/QcOfpuPGnk09\nmhcZIbI5WSYfsPQbXpu02mnmanXgl7Pt4teQGdQBTq0cO4xj7FOdEIcxD8WAmF9E\ntS+DhrNoOaaR2mwicHZ5+cfMKviNWw3qmolDWZKm8tIZTU21+lU+LyB6uMwCgQ/y\nCZG4eY0O+CKlzn6nlsOLgpKhZDtiNgPN3eGcIHfw+oxCECH6de/Ewuk3Co6Thnh1\nRu6Lt+g/od2IYvKl0R6efBTeuwozoubJ\n-----END CERTIFICATE-----\n",
"key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA5L/rjXT0dEBmN3BSnsmLCwF9goOZWb5RNvvc+dnUQnaG5nJ8\nxMPSpEAm4y2ZlC8Uhm56H+7A2cUf/ogkUcKkyQHyXQrjjPzec5h3+734MmitvHAA\nzC/j9smDOvdde9evMUDH9Sn7oJtpz+MKckTFvsBoLGuBPiuLn8P9HCmLKlMNjR2h\nfJab+DZG+eysAGkoiGS9zOxMVF9Xsvk/EHxm5p3ZPF2oOEr8cimgQO/ew29+fF4e\ntwskre/AA8KCw/6Bg2XwZ3HmFPJ9nd/zJSzqm5U96WFEth4ABVqZE4MS4rRo2K+e\nQy2rXzDnmiL3cGiLuL9aal+9JCFRxKE1RSIdEQIDAQABAoIBAHmnb/itKIzi6vm0\n7NuxyBa0VjGhF19ZDgw16pGePXqTWq8YWC61DkN4MrZDPBhI6ZuNCboN2dZ3NcrC\nUL6Cy+xy8ph1AAutOAk2HyltIKB+d1duIZ52IcDP7tDfWYJRdMS29SD3kPEbdiyv\nTJD07k3COiTVj8imk/0F2Iivt2lr/GLyPSKFckNzCC3Tfz7pCaq2+JBxt08QtZPl\nqL81or7+BpcUNgmZyA1WLZsMpoYeXX6fsndUrVcfZIayaIXjX9NJYxQuwW44EYxq\nhBRS+goxXo+xQSKCSEuPf36pc82sxUiNEszAOd5Iqjrl1pRVVoLFdVLXFaQ0f0J8\nq7PXjYECgYEA/mhXuoTQuVsJvqCfZSLs/bZnJotSi9na4oOmZBp+LokRzHYbmchm\nSjX1scc9fm5MMOe82WAUMBYZhBVJrjqk/rQvGBMpRqut6JiujpSDR27xAcpJD79P\ngBre3PSDJ84koI8PTyGTcPTa3RqY5+5Iv4JsFPFCshFGMI/ldwRxickCgYEA5i52\nuSyMohMrAfHyS3IG1c/bGrhSZ7O790nUakEGpwi8paD/qi4klb5LYmJLQbaZOKes\nh6voXQ3Zhvd0j6sXzi2d802OlB9LIq+FcctiiS3wR1YykbAq3RDAfYMEEnSdRnp3\noz8i3l1y/a1bP3RnjSmeaMVG/nITFlfMbfCEnQkCgYBemQvt9g7qrVhdQrqiT69R\n0+5dHbcu+23xhkRrupIq2Zr9rPksYKDwfUoDtfM+vOKl2LWXGqvHCaCpRYUlPPc3\nImbUi+NwPMwozgUyTTTXbgA9yysJqPh1yQgPnvfZ6EQkU628nd6GRPXQ+1/Z9fel\nBmkMDH3hWpz/17HaZJOXSQKBgETamUEDBn5k5XSLf0L6NPk4V/5CLMRAi3WJbDTs\nhqTohCW3Z0Ls0pzIc5xWctSRXnwIDB/5WGSdg/hPhVqEf3Z5RspE5OWCBuO1RWGo\nySznxPxR2Iaj/+5o2GuzCUDMCU/PyoHWnQOPSJqBhM4Sb/dV/8CvYnEyhmskkE5C\nqCihAoGBAJr0fISYmFtX36HurmFl+uCAcOPXy6qEX3Im6BA6GiNGmByk+qFVP5e9\nLRjX7LouGlWyPridkODdWAwCTxn1T7FcGFRZlK0Ns3l8Bp+FMznXX05TlkNZpC6X\nV9F0GVQtrwsos605Z8OG4iMRkn+3Qk6FZ8QnlBX807killRcGlC4\n-----END RSA PRIVATE KEY-----\n"
}' --json-input
Description:
Adds a certificate to be used in SSO configuration. Certificate data have to be line delimited with \n character and may contain —–BEGIN and —–END comments. Refer to certificate later in configuration with <name>.key for key and <name>.pem for certificate. After
every change in certificate store SSO service must be restarted. To delete a certificate use ntk cfg delc sso.certificates.server
. Refer to these certificates as “server.pem” and “server.key” elsewhere in configuration, where server is the unique certificate name.
Enabling modules
Typical module configuration
$ ntk cfg setc :sso.modules '[ "cron", "startpage", "ldap", "notakey" ]' --json-input
Description:
Configure a list of enabled modules. User repositories and authentication filters define which modules are required. For basic setup with LDAP (or Active Directory)
the list above is sufficient. Follow each change with ntk sso restart
.
Plugins
It is possible to configure additional custom modules for NtkSSO to customize authentication workflow or add custom templates. All plugins must be bundled as docker images that provide code and template files to the main container. See Notakey Examples for skeleton project that can be used to build plugin images.
To configure plugin add plugin to modules array in format [module name]:[repository path]:[image tag].
$ ntk cfg setc :sso.modules '[ "cron", "startpage", "ldap", "notakey", "mytheme:myorg/sso-mytheme:latest" ]' --json-input
Adding service provider logos
Use scp to copy logo files to /home/ntkadmin/userlogos directory. Files can be in any format, PNG with transparent background preferred, use 250x80px for optimal scaling.
$ scp mylogo.png ntkadmin@ntk.example.com:/home/ntkadmin/userlogos/
You can use URL https://sso.example.com/userlogos/mylogo.png later as startpage logo for required SP.
Command reference
ntk sso restart
Usage:
ntk sso restart
Description:
Stops and starts SSO service. Use this command to apply configuration changes.
ntk sso start
Usage:
ntk sso start [--force] [--update]
Description:
Starts SSO service. Use –force option to skip TLS certificate validation matching configured DNS hostname e.g. useful during initial setup. Use –update to update custom plugin modules, if any configured. If specified, new images will be loaded from configured remote repository.
ntk sso stop
Usage:
ntk sso stop
Description:
Stops SSO service. Service will be stopped and will not be started until ntk sso start
is called. The
stopped state persists across appliance restarts.
ntk sso validate
Usage:
ntk sso validate
Description:
Validates SSO service configuration.
ntk sso status
Usage:
ntk sso status
Description:
Shows status of SSO service.
ntk sso sync
Usage:
ntk sso sync [--delete-missing]
Description:
Pulls newer content and publishes any locally added or edited files to cluster datastore. Location of local directory is defined in /opt/notakey/cfg/.env as ASSETS_DIR. Usually this is /home/ntkadmin/userlogos directory. Adding –delete-missing removes missing files that are stored in cluster, but missing locally. Any files exceeding 350kB in size will be ignored and must be copied manually to other nodes.
Authentication Proxy
Commands in ntk ap
section control Notakey Authentication Proxy.
Configuration settings
The following configuration settings must be present to start this service.
Key | Description | Default | Possible values |
---|---|---|---|
:ap.vpn_port_in | Incoming RADIUS service port that the Notakey Authentication Proxy listens for inbound requests from VPN access gateways. | 1812 | 1-65000 |
:ap.vpn_port_out | RADIUS server port where Notakey Authentication Proxy will send upstream requests. | 1812 | 1-65000 |
:ap.vpn_radius_address | RADIUS server hostname or IP address that Notakey Authentication Proxy forwards the requests. Specify multiple hosts by seperating them by :, e.g. “192.168.0.1:192.168.0.2” | (none) | string |
:ap.vpn_secret_in | CHAP secret for incoming requests from VPN access gateways. | (none) | string, max 16 characters |
:ap.vpn_secret_out | CHAP secret for outgoing requests to the RADIUS server. | (none) | string, max 16 characters |
:ap.vpn_access_id | Application Access ID as seen on Dashboard Application settings page. | (none) | string |
:ap.vpn_message_ttl | Authentication message time to live time in seconds. Should match the RADIUS timeout value of VPN gateway. | 300 | integer |
:ap.message_title | Authentication message title. Title is shown in application inbox for each authentication message, e.g. “VPN login to example.com” | “VPN authentication” | string |
:ap.message_description | Authentication message contents. Message is shown in push notification and inbox. Use {0} as a placeholder for username. | “Allow user {0} VPN login?” | string |
:ap.proxy_receive_timeout | RADIUS message internal receive timeout. | 5 | integer |
:ap.proxy_send_timeout | RADIUS message internal send timeout. | 5 | integer |
:ap.api_provider | Backend API provider. Do not configure if using local instance. URL needs to be in format https://host/api | <local instance> | string |
:ap.api_timeout | Backend API call request timeout. | 5 | integer |
:ap.cisco_acl_passthrough | Cisco ownloadable ACL support, when using Cisco ACS as RADIUS server | false | true,false |
:ap.custom_params | Key value pairs of custom configuration options (cluster value with fallback to local node value). | (none) | key1=value,key2=value |
Change settings using ntk cfg setc
command.
Command reference
ntk ap start
Usage:
ntk ap start
Description:
Starts Notakey Authentication Proxy.
ntk ap stop
Usage:
ntk ap stop
Description:
Stops Notakey Authentication Proxy. Service will be stopped and will not be started until ntk ap start
is called. The
stopped state persists across appliance restarts.
ntk ap status
Usage:
ntk ap status
Description:
Shows status of Notakey Authentication Proxy container instance.
ntk ap validate
Usage:
ntk ap validate
Description:
Validates Notakey Authentication Proxy configuration, checks port, RADIUS server and CHAP secret configuration, validates service Access ID.
CLI Management Utility
Commands in section ntk cli
simplify management tool updates and automate install of new versions.
Command reference
ntk cli check
Usage:
ntk cli check
Description:
Checks if new version of configuration utility is available.
ntk cli tag
Usage:
ntk cli tag
ntk cli tag <new tag>
Description:
Set tag of configuration utility. Tag configuration represents an update channel. Use this option to control the update flow and load new functionality. Updates in single tag include fixes for current functionality, but do not contain new functionality. Please consult your support representative for compatibility of your appliance with non-standard tag versions.
ntk cli update
Usage:
ntk cli update
Description:
Load the latest release of configuration utility respecting the tag (update channel).
Networking Configuration
To configure network settings, you must use the ntk net
command.
The general syntax is net ntk <interface> <command>
, where interface
is e.g. eth0
.
Command reference
ntk net hostname
Usage:
ntk net hostname
ntk net hostname <new hostname>
Description:
Displays or sets the Notakey Authentication Appliance hostname. The hostname will be also used to identify node in cluster environment. Hostname cannot contain dots or any special symbols.
ntk net domain
Usage:
ntk net domain
ntk net domain <new domain>
Description:
Displays or sets the Notakey Authentication Appliance domain. The domain will be used to group nodes in cluster environment. Nodes in single L2 network should share domain value. Domain value cannot contain dots or any special symbols.
ntk net address
Usage:
ntk net eth0 address
ntk net eth0 address <ip address/cidr mask>
Description:
Type ntk net <interface> address
to see the appliance’s IP addresses.
Type ntk net <interface> address <ip_in_cidr>
to set the IP address for the
specified network interface.
The IP address must be in CIDR notation - e.g. 10.0.1.5/24
.
Please use IP subnet calculator to convert netmask to subnet mask bits. Changes are applied immediately, so in case of SSH connection ensure that
ntk net dhcp
Usage:
ntk net eth0 dhcp
ntk net eth0 dhcp on
ntk net eth0 dhcp off
Description:
Type ntk net <interface> dhcp
to see if the specified interface is using DHCP.
To turn DHCP on or off, type ntk net <interface> dhcp [on|off]
, e.g.
ntk net eth0 dhcp off
to disable DHCP.
ntk net gateway
Usage:
ntk net eth0 gateway
ntk net eth0 gateway <gateway ip>
Description:
Type ntk net <interface> gateway
to see the interface’s gateway.
Type ntk net <interface> gateway <ip>
to set the interface’s gateway. E.g. ntk net eth0 gateway 10.0.1.1
.
ntk net dns
Usage:
ntk net dns
ntk net dns <server1,server2>
Description:
Display or configure one or more DNS servers. Separate multiple IP addresses by (,) comma. Changes are applied immediately.
ntk net proxy
Usage:
ntk net proxy
ntk net proxy on
ntk net proxy off
Description:
Enable or disable proxy server for outbound connections. Server can be configured with ntk net proxy_host
command.
Change is applied after full appliance restart.
ntk net proxy_host
Usage:
ntk net proxy_host
ntk net proxy_host <http[s]://proxy-host:port>
Description:
Display or set the proxy server used for outbound HTTP/HTTPS connections. Use https:// prefix if TLS is used on proxy server.
Backup and Restore
Command in backup section provide interface for saving and restoring appliance configuration to and from file.
Command reference
ntk backup save
Usage:
ntk backup save <path to file>
Description:
Save database and configuration backup to file. Files are usually located in user’s home directory e.g. /home/ntkadmin for persistent storage or in /tmp for temporary location. Utility scp can be used to transfer files between nodes or to backup location.
ntk backup load
Usage:
ntk backup load <path to file>
ntk backup load --skip-net <path to file>
ntk backup load --skip-cluster <path to file>
Description:
Restore configuration from backup. Use –skip-net to avoid locally set network configuration overwrite. Appliance must be fully restarted for changes to be applied and all services manually started after restart. Use –skip-cluster to skip configuration and data restore for cluster wide configuration. This is needed if doing in-place upgrades. Both options –skip-cluster and –skip-net can be combined.
VRRP Configuration
The appliance has built-in support for the Virtual Router Redundancy Protocol (VRRP) protocol. VRRP provides functionality to share single IP address
between multiple nodes in cluster and it can be managed with the ntk vrrp
command line utility.
Configuration allows to define multiple shared addresses (called instances) and bind them to defined resource availability trackers.
Every instance can be in three states:
MASTER - when serving the shared IP address
FAULT - when one of bound trackers fails
BACKUP - when all trackers return OK status, but another MASTER instance is active for this resource (IP address).
Commands below configure the default instance called nas, to configure additional instances use ntk cfg set vrrp.services.*
commands.
Available subcommands are:
Command | Description |
---|---|
address | Get/set the virtual IP address |
interface | Get/set the network interface |
peers | Get/set the list of physical unicast peers |
status | See VRRP status |
vrid | Get the virtual router ID |
start | Enable the VRRP service |
stop | Stop the VRRP service |
track | Define bound trackers to this default instance |
validate | Check config for errors |
preempt | Define preempt mode |
prio | Configure local instance priority |
Command reference
ntk vrrp address
Usage:
ntk vrrp address
ntk vrrp address <server ip>
Description:
Type ntk vrrp address
to see the configured virtual IP address.
Type ntk vrrp address <address>
to set the configured virtual IP address. It must be in
decimal notation - aaa.bbb.ccc.ddd
.
ntk vrrp track
Usage:
ntk vrrp track
ntk vrrp track set disk,nas,sso,generic,port...
ntk vrrp track clear
Description:
Type ntk vrrp track
to see the configured trackers for default VRRP instance.
Type ntk vrrp track set <comma_separated_tracker_list>
to set the available trackers.
See section “Tracker configuration” for more details.
ntk vrrp interface
Usage:
ntk vrrp interface
ntk vrrp interface eth0
Description:
Type ntk vrrp interface
to see the configured network interface.
Type ntk vrrp interface <iface>
to set the network interface. It must match one
of the detected network interfaces on the appliance (e.g. eth0
).
ntk vrrp peers
Usage:
ntk vrrp peers
ntk vrrp peers clear
ntk vrrp peers set 10.0.1.5,10.0.1.6
Description:
Type ntk vrrp peers
to see the list of unicast peers.
Type ntk vrrp peers clear
remove all peers from the list.
Type ntk vrrp peers set <peer_list>
to set a comma-separated peer list. E.g. 10.0.1.5, 10.0.1.6
.
By default multicast is used for communication between nodes. If no multicast is available on transport layer please configure all cluster members using this command.
ntk vrrp status
Usage:
ntk vrrp status
Description:
Type ntk vrrp status
to see the status of all VRRP instances.
$ VRRP status:
status: running
Instance status:
Instance Address Status In status since
ap 10.0.1.99 BACKUP Mon Jun 3 11:43:33 EEST 2019
nas 10.0.1.50 BACKUP Mon Jun 3 11:43:33 EEST 2019
ntk vrrp vrid
Usage:
ntk vrrp vrid
ntk vrrp vrid 51
Description:
Type ntk vrrp vrid
to get the virtual router ID.
Type ntk vrrp vrid <vrrp ID>
to set the virtual router ID.
Nodes providing specific VRRP service must share the same vrid. Avoid using this value on other services as this will cause interference, especially if used in the same physical network.
Default value for virtual router ID is 51.
ntk vrrp start
Usage:
ntk vrrp start
Description:
Type ntk vrrp start
to start the VRRP service.
ntk vrrp stop
Usage:
ntk vrrp stop
Description:
Type ntk vrrp stop
to stop the VRRP service.
ntk vrrp restart
Usage:
ntk vrrp restart
Description:
Type ntk vrrp restart
to restart the VRRP service.
ntk vrrp preempt
Usage:
ntk vrrp preempt
ntk vrrp preempt on
ntk vrrp preempt off
Description:
Type ntk vrrp preempt
to see if the preempt mode is enabled.
Type ntk vrrp preempt on
enable preempt mode.
Type ntk vrrp preempt off
disable preempt mode.
Preempt mode allows to guarantee that service will be served by the server with the highest priority when cluster is fully functional. Preempt mode is disabled by default and the service will remain with the last working node, even if new nodes appear online with higher priority.
ntk vrrp prio
Usage:
ntk vrrp prio
ntk vrrp prio 100
Description:
Type ntk vrrp prio
to get the node priority.
Type ntk vrrp prio <prio>
to set the node priority.
Priority controls the service takeover conditions with respect to preempt setting. If active node fails, service will be taken over by a node with next highest prio value. To guarantee preempt takeover when all nodes ar functional, priority difference has to be at least 50.
General VRRP options
Configuration option | Default | Description |
---|---|---|
vrrp.syslog | - | IP address or hostname to send VRRP event logs. Events will be tagged with tag vrrp and user facility. |
Tracker configuration
VRRP includes several bundled trackers to monitor services and resources on local node. Each tracker has configuration
block defined with ntk cfg set vrrp.track.<tracker name>
.
Available trackers:
Tracker id | Description |
---|---|
disk | Monitors that disk usage is below certain treshold |
nas | Notakey Authentication server specific tracker |
sso | SSO server specific tracker |
rp | Reverse proxy service tracker |
port | Monitor that local TCP/IP socket is open |
tcp | Monitor that remote or local TCP/IP socket is open |
generic | Monitor user defined HTTP(s) URL for availability (must return 2xx status) |
Configuration option | Default | Scope (tracker name) | Description |
---|---|---|---|
vrrp.track.xxx.interval | 5 | Global | Interval in seconds between runs of this tracker |
vrrp.track.xxx.timeout | 4 | Global | Timeout in seconds for tracker to wait for response (supported by nas, sso, rp, generic) |
vrrp.track.xxx.fall_count | 2 | Global | Number of failures required to put this instance in FAULT state |
vrrp.track.xxx.rise_count | 2 | Global | Number of sequential successful executions required to put this instance in BACKUP or MASTER state |
vrrp.track.port.subject | - | port | Port and protocol to monitor, e.g. tcp/80 or udp/1812 |
vrrp.track.generic.url | - | generic | Url for monitoring availability |
vrrp.track.disk.treshold | 98 | disk | Percentage of disk usage up to which usage is considered normal |
Additional VRRP instance definition
Below is configuration options for every VRRP instance.
Configuration option | Default | Description |
---|---|---|
vrrp.services.xxx.prio | - | Same as ntk vrrp prio |
vrrp.services.xxx.vrid | - | Same as ntk vrrp vrid |
vrrp.services.xxx.interface | - | Same as ntk vrrp interface |
vrrp.services.xxx.address | - | Same as ntk vrrp address |
vrrp.services.xxx.preempt | off | Same as ntk vrrp preempt |
vrrp.services.xxx.preempt_delay | 0 | Configure delay in seconds before transitioning to lower prio node |
vrrp.services.xxx.track | - | Same as ntk vrrp track |
vrrp.services.xxx.secret | - | Configure password protection for this instance |
vrrp.services.xxx.pulse_interval | 1 | Seconds between heartbeat messages from MASTER to BACKUP instances |
Below is an example of authentication proxy VRRP instance configuration.
# Define additional VRRP instance to run authentication proxy
# and monitor that the RADIUS udp socket is accessible.
ntk cfg set vrrp.services.ap '{
"prio": "100",
"vrid": "52",
"interface": "eth0",
"address": "10.0.1.99",
"preempt": "off",
"track": [
"port", "nas"
]
}' --json-input
ntk cfg set vrrp.track.port '{
"interval": "20",
"subject": "udp/1812"
}' --json-input
Cluster Management
When Notakey Authentication Appliance is deployed in cluster environment, commands in ntk cluster
allow individual node state
management in cluster and provides cluster health overview.
Command reference
ntk cluster start
Usage:
ntk cluster start [--if-not-running]
Description:
Starts local cluster member datastore node. Node can work standalone or can be joined to other nodes, if datacenter settings are
consistent. Use conditional start --if-not-running
to avoid error if datastore is already started.
ntk cluster stop
Usage:
ntk cluster stop
Description:
Stops local cluster member datastore node. If after stop cluster no longer holds quorum, this command can lead to read-only datastore on other nodes, use with care.
ntk cluster restart
Usage:
ntk cluster restart
Description:
Restarts local cluster member datastore node. This command can lead to read-only datastore on other nodes, use with care.
ntk cluster status
Usage:
ntk cluster status
Description:
Shows datastore status.
ntk cluster validate
Usage:
ntk cluster validate
Description:
Validates datastore configuration settings. Checks include node name validity, datacenter and cluster IP configuration.
ntk cluster cleardata
Usage:
ntk cluster cleardata
Description:
This command is intended for use, where cluster leader/follower data stores can not be reconciled. Clear the data on followers, and allow cluster leader to replicate to them.
Overview of cluster rebuild steps :
Issue
ntk cluster leave
on nodes out of sync.Use
ntk cluster cleardata
to purge all local data.Use
ntk cluster reset
to purge data about remote nodes.Restart appliance to purge any in-memory saved state.
Issue
ntk cluster join ...
to join to root node.Verify that cluster is in sync with
ntk cluster nodes
ntk cluster join
Usage:
ntk cluster join <ip address of root node>
Description:
Joins existing cluster node and starts data replication. ACME certificates will also be replicated during join.
ntk cluster health
Usage:
ntk cluster health
Description:
Displays health information about the cluster nodes and datastore replication status. A warning information will be shown if cluster is in degraded state. RAFT and internal protocol information is required during upgrades as different versions supporting different sets of versions cannot coexist in the same cluster.
$ ntk cluster health
=> Cluster Datastore status
RAFT version: v3
protocol version: v2
Leader: ntknode1
=> Cluster members
Node IP address Status Member ID
ntknode1 172.16.112.136 alive 47804985-5aab-2e37-6c87-b60aa22d3925
ntknode2 172.16.112.137 alive 23423985-34ab-2e37-6123-b60aa22d4524
ntknode3 172.16.112.138 alive 34535985-123b-2e37-1287-b60aa22d3932
ntk cluster myid
Usage:
ntk cluster myid
Description:
Shows persistent ID number of member node. This information is relevant for troubleshooting when reassembling cluster after general failure.
ntk cluster snapshot
Usage:
ntk cluster snapshot </absolute/path/to/file>
Description:
Creates a cluster datastore snapshot thet can be used to recover datastore after failure or can be used for low-level appliance inspection.
ntk cluster leave
Usage:
ntk cluster leave
Description:
Informs other cluster members about leave intent and removes reference data for cluster. Cluster will remain functional and other nodes will continue to serve requests. This command is not recommended when leave candidate node is in active state serving requests, as data loss is imminent. Please use VRRP configuration to demote node from active state and then issue leave command.
ntk cluster nodes
Usage:
ntk cluster nodes
Description:
Provides overview about known nodes and their state in cluster. Use this command to identify failures in cluster mode.
$ ntk cluster nodes
Node Status IP Address Data Center
as-1 alive 10.1.1.62 master-dc
as-4 alive 10.1.1.61 master-dc
as-3 alive 10.1.1.65 master-dc
as-2 alive 10.1.1.66 master-dc
ntk cluster reset
Usage:
ntk cluster reset <comma_separated_peer_list>
Description:
Use this command for cluster recovery, if quorum has been lost due to failed nodes. Pass a comma-separated peer list, and a new cluster will be initialized, inheriting data from the previous cluster.
Steps to take for recovery:
Run this command on every node specified in the peer list
On every node execute ‘ntk as restart’
Built-in Troubleshooter
Use doctor command to validate your configuration.
Command reference
ntk doctor
Usage:
ntk doctor
Description:
Run ntk doctor
to attempt identify common problems with
configuration, and suggest solutions.
Setup Wizard
Issue ntk wizard
on every node to configure the initial settings to get applicance connected to network and acquire TLS
certificates.
Command reference
ntk wizard
Usage:
ntk wizard
Description:
Run ntk wizard
to setup initial configuration required to start Notakey Authentication Server service in a single node
mode.
Wizard will require networking, external FQDN and TLS configuration.
Configuration can then be later adjusted using specific commands.
System Maintenance
Commands in ntk sys
control Notakey Authentication Server state and allows maintenance operations.
Command reference
ntk sys passwd
Usage:
ntk sys passwd
Description:
Prompts user for new password. Use this command to change password after appliance image deployment.
ntk sys reboot
Usage:
ntk sys reboot
Description:
Requests an immediate appliance restart.
ntk sys shutdown
Usage:
ntk sys shutdown
Description:
Shuts down appliance.
ntk sys vmtools
Usage:
ntk sys vmtools enable
ntk sys vmtools disable
Description:
Enables or disables VMware tools service. Must be enabled only in VMware based deployment e.g. when using VMware ESXi. Tools installation allows external power state control of appliance OS and provides diagnostic and networking information to hypervisor.
ntk sys docker
Usage:
ntk sys docker
ntk sys docker 17.04.0-ce
Description:
$ ntk sys docker
Refreshing package state...ok
Current version: 17.05.0
Available versions:
1.10.3
1.11.2
1.12.6
1.13.1
17.03.1-ce
17.04.0-ce
17.05.0-ce
$ ntk sys docker 17.05.0-ce
Refreshing package state...ok
Downloading install image...ok
Changing docker version to 17.05.0-ce...ok
Please reboot appliance with "ntk sys reboot" to apply docker engine switch.
All services on NtkAA node are deployed as docker container. Sometimes it is neccessary to change docker. Call command without arguments to discover the available docker engine versions. To change version pass new version as the argument. Available versions are loaded from repository URL configured with “ntk sys repo” command. Use this command with care as changes in docker version can affect all user services. A full system restart is required after version switch.
ntk sys ntp
Usage:
ntk sys ntp <server1,server2>
ntk sys ntp
Description:
Configure or verify NTP time servers. NTP service uses unicast UDP/123 for communication, please ensure the required port is open in firewall.
ntk sys df
Usage:
ntk sys df
Description:
Shows free disk space and usage by volumes and containers.
ntk sys growfs
Usage:
ntk sys growfs
Description:
Resizes (increase only) root filesystem during startup to allocate all available storage. You will need to use “ntk sys reboot” to restart node. Run this command every time hardware size is increased, as resize will happen only once during next restart.
ntk sys arp
Usage:
ntk sys arp
Description:
Check local IPv4 MAC arp table. Useful for network connectivity validation.
ntk sys ping
Usage:
ntk sys ping <host>
Description:
Ping specified host with ICMP packets. Useful for network connectivity validation.
ntk sys traceroute
Usage:
ntk sys traceroute <host>
Description:
Check IP routing path to specified host. Useful for network connectivity validation.
ntk sys perf
Usage:
ntk sys perf
Description:
Check overall system load and memory usage. Shows list of most active processes.
ntk sys repo
Usage:
ntk sys repo https://index.notakey.com/notakey-support/appliance/raw/v<version> --auth <user:pass>
ntk sys repo
Description:
Configure source repository for update status check and new package download.
Default value is https://index.notakey.com/notakey-support/appliance/raw/v<version>
where <version> is the installed appliance version.
In most cases, when upgrading within single major version, an upgrade is possible by adjusting this value to required version.
Any upgrades should be exercised with care, while cluster host is in idle state.
$ ntk sys repo https://index.notakey.com/notakey-support/appliance/raw/v5.0.0
==> Changing repository to "https://index.notakey.com/notakey-support/appliance/raw/v5.0.0"
Testing service connectivity... ok
Checking index... ok
Clearing cache... ok
Updating configuration... ok
Configuration complete, use "ntk sys update" to fetch new package versions.
ntk sys update
Usage:
ntk sys update [--force]
Description:
Retrieve list of latest packages, perform update if new versions are found. Use --force
to reinstall all locally downloaded packages.
$ ntk sys update --force
==> The following package versions will be downloaded
redownloading all available packages
dashboard: dashboard:2.7.1
auth-proxy: authproxy:1.0.2
CLI: cfg:4.1.0
consul: consul:0.8.4
traefik: traefik:1.3.1
vrrp: keepalived:1.3.9
htop: htop:latest
Do you want to proceed? (y/n): y
==> Downloading from repo.notakey.com
downloading dashboard:2.7.1
downloading authproxy:1.0.2
downloading cfg:4.1.0
downloading consul:0.8.4
downloading traefik:1.3.1
downloading keepalived:1.3.9
downloading htop:latest
==> Configuration update
apply new versions after service restart? (y/n): y
restart all running services? (y/n): y
=> Graceful shutdown of existing instances
...
update complete
ntk sys cleanup
Usage:
ntk sys cleanup [--all-unused]
Description:
Remove unused packages from local repository. If --all-unused
is specified, also images with stopped containers are deleted, otherwise only unused images are removed. Note that this is a disk IO intensive process and should be used on a lightly loaded servers, as service could be impacted.
ntk sys supreq [–include-config] [–encrypt]
Usage:
ntk sys supreq
Description:
Create an archive with all logfiles and diagnostic information for support request submission. No user data is contained in exported archive, except authentication history and logs. Archive contains sensitive configuration details, sharing with third parties should be exercised with caution. Expect a significant IO load increase while report is running. If –include-config is specified, report archive will contain your LOCAL node configuration. Local configuration usually contains your TLS certificate private keys, RADIUS access secrets, etc. If –encrypt is specified, you will be prompted for password and file will be encrypted with 256-bit AES ecryption with Cipher Block Chaining.
$ ntk sys supreq
=> Gathering system detail... ok
=> Processing config... ok
=> Gathering logs... ok
=> Gathering system service details... ok
=> Gathering application details... ok
=> Packing... ok
Support request archive created in /home/ntkadmin/support_request_20170628135340.tgz
ntk sys timezone
Usage:
ntk sys timezone
ntk sys timezone <timezone>
Description:
Set timezone for containers. Timezone is in zoneinfo format, e.g. UTC, Europe/Riga or America/New_York. All running services will need to be restarted to apply timezone change. This affects time display in applications and log display time values.
Plugin support
Appliance supports running of both managed or third party container images to support your requirements. Managed plugins are the ones published by Notakey Latvia on
our container repository. Managed plugins are updated automatically with command ntk sys update
when new releases appear and typically are backwards compatible with your infrastructure.
Third party plugin can be any container image and typical use cases include LDAP (OpenLDAP) for storing user repository, SAML supporting reverse proxy for third party web application protection
(see example here), OpenID Connect provider for custom onboarding requirements and many more.
Running plugins depends on cluster service, this can be started with ntk cluster start
, please consult Cluster Management section for more details.
Plugin configuration is global on all nodes in cluster group, so any configuration changes applied to one node affect all other nodes in the same group. To apply configuration changes a ntk plugins restart
command must be run on all nodes.
Configuration settings
Key | Type | Description |
---|---|---|
:plugins.?.ports | array | Defines one or more ports to expose externally directly on node. Use external-port:internal-port pattern if external port is not the same as internal container port. |
:plugins.?.image | string | Set image to load for this plugin. Not required for managed plugins (see below). |
:plugins.?.tag | string | Defines a managed plugin by Notakey Latvia. Tag names are provided for specific integration projects. |
:plugins.?.env | struct | Environment variables to pass to this container. This is a key-value mapping. |
:plugins.?.volumes | array | List of volumes to expose to this container. |
:plugins.?.host | string | External FQDN for http service provided by container that will be exposed using HTTP using builtin reverse proxy. Internal port must be defined, if using host. |
:plugins.?.internal_port | integer | Internal port of container that provides http service towards builtin reverse proxy. |
Examples
Expose a plugin using reverse proxy
Example below runs Apache2 with specific configuration and makes it accessible externally on https://mysp.example.com.
# Expose http service externally with FQDN mysp.example.com
# If Let's Encrypt ACME client is eanbled, HTTPS TLS certificate will be acquired automatically
$ ntk cfg setc :plugins.spproxy.host "mysp.example.com"
# Plugin container exposes port 9000 for plain HTTP requests
$ ntk cfg setc :plugins.spproxy.internal_port 9000
# Define image to load for this plugin
$ ntk cfg setc :plugins.spproxy.image "soarinferret/mod_auth_mellon:debian"
# Define volumes for this container
# /home/ntkadmin/shared/saml.conf is the path on host VM, /usr/local/apache2/conf/httpd.conf is internal container path
# ro means that map is read only and cannot be modified by container
$ ntk cfg setc :plugins.spproxy.volumes[0] "/home/ntkadmin/shared/saml.conf:/usr/local/apache2/conf/httpd.conf:ro"
# maps another readonly volume /home/ntkadmin/shared/mellon to /mellon
$ ntk cfg setc :plugins.spproxy.volumes[1] "/home/ntkadmin/shared/mellon:/mellon:ro"
# Pull plugins from hub.docker.com
$ ntk plugins update
# Start plugin services
$ ntk plugins start
Internal service exposed directly on node
Example below runs a LDAP server on appliance node and makes it accessible on standard ports tcp/389 and tcp/636 (LDAPs).
# Define external ports for this plugin
$ ntk cfg setc :plugins.ldap.ports[0] "389:389"
$ ntk cfg setc :plugins.ldap.ports[1] "636:636"
# Define image to load for this plugin
$ ntk cfg setc :plugins.ldap.image "osixia/openldap:latest"
# Define environment variables for this container
$ ntk cfg setc :plugins.ldap.env.LDAP_ORGANISATION "Internet Widgets, Ltd"
$ ntk cfg setc :plugins.ldap.env.LDAP_DOMAIN "company.com"
# Define volume for LDAP database files
$ docker volume create ldap-data
$ ntk cfg setc :plugins.ldap.volumes[0] "ldap-data:/var/lib/ldap"
# Define volume for LDAP config files
$ docker volume create ldap-config
$ ntk cfg setc :plugins.ldap.volumes[1] "ldap-config:/etc/ldap/slapd.d"
# Pull plugins from hub.docker.com
$ ntk plugins update
# Start plugin services
$ ntk plugins start
Command reference
ntk plugins start
Usage:
ntk plugins start [--force]
Description:
Starts all configured plugins. Use –force option to skip configuration validation.
ntk plugins stop
Usage:
ntk plugins stop
Description:
Stops all configured and started plugins. Service will be stopped and will not be started until ntk plugins start
is called. The
stopped state persists across appliance restarts.
ntk plugins validate
Usage:
ntk plugins validate
Description:
Validates all plugin configuration.
ntk plugins status
Usage:
ntk plugins status
Description:
Shows status of all configured plugins.
ntk plugins restart
Usage:
ntk plugins restart
Description:
Stops and starts all configured plugins. Use this command to apply configuration changes.
ntk plugins update
Usage:
ntk plugins update
Description:
Updates all plugin container images. If this is a managed plugin (has tag
configured)
appliance will perform a metadata lookup to see if there is a new release. If new release is found,
it will be pulled and configuration updated with new image version information.
For third party plugins the configured image will be pulled to see if there are any updates.
Image information will not be changed. To change image information use ntk cfg setc :plugins.?.image <image>
command.
Zabbix Agent
The appliance has built-in support for Zabbix agent to be connected to Zabbix server or proxy.
Configuration settings
Key | Description | Default | Possible values |
---|---|---|---|
zabbix.listen_port | Local listen port of Zabbix agent. | 10050 | 1-65000 |
zabbix.server_fqdn | Specify remote Zabbix server or proxy IP address or hostname. | (none) | - |
zabbix.server_port | Remote Zabbix server or proxy port. | 10050 | 1-65000 |
zabbix.server_version | Server major.minor version, e.g. 4.2 . Must be present to load correct agent version. |
(none) | - |
zabbix.debug_level | Debug level as defined by agent container 1 - least debugging, 5 - max debugging. | 1-5 | 1 |
zabbix.custom_params | Key value pairs of custom configuration options (local, non-cluster value), e.g. key1=value,key2=value . |
(none) | - |
Change settings using ntk cfg setc
command.
Basic setup
# Configure remote server IP
$ ntk cfg set zabbix.server_fqdn zabbix.example.com
# Set server version you are using
$ ntk cfg set zabbix.server_version 4.4
# Load matching agent version
$ ntk sys update
==> The following package versions will be downloaded
downloading only new packages
zabbix-agent: notakey/zabbix-agent:4.4-latest
Do you want to proceed? (y/n): y
==> Downloading from repo-test.notakey.com
downloading notakey/zabbix-agent:4.4-latest
==> Configuration update
apply new versions after service restart? (y/n): y
restart all updated services? (y/n): y
# Start agent
$ ntk zbx start
=> Validating Zabbix Agent configuration
Checking server FQDN... ok
Checking server version... ok
Checking software image...
required: notakey/zabbix-agent:4.4-latest
Checking listen port... ok
Checking server port... ok
=> Graceful shutdown of existing instance
shutting down Zabbix Agent instance ... ok (shut down 7296da5867bdbe26366fa6d669455d48dbf87d8dea6d0eb2e140daae2ed9cd8e)
=> Starting new Zabbix Agent instance
server FQDN: zabbix.example.com
server port: 10051
server version: 4.4
listen port: 10050
image: notakey/zabbix-agent:4.4-latest
timezone: Europe/Riga
started 8a4e65ec9bd5a7df2d8d605fac594e064640774e222976dd7bd4fa2b3fb01941
# Inspect status with docker logs command
$ docker logs zabbix-agent
** Deploying Zabbix agent
** Preparing the system
** Preparing Zabbix agent
** Preparing Zabbix agent configuration file
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "PidFile": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "LogType": 'console'... added
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "LogFile": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "LogFileSize": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "DebugLevel": '1'... added
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "SourceIP": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "EnableRemoteCommands": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "LogRemoteCommands": ''... removed
** Using 'zabbix.example.com' servers for passive checks
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "Server": 'zabbix'... updated
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "ListenPort": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "ListenIP": ''... removed
** Updating '/etc/zabbix/zabbix_agentd.conf' parameter "StartAgents": ''... removed
** Using 'zabbix.example.com:10051' servers for active checks
.....
Updating Zabbix server
If server version is changed, agent version may need to be adjusted to support newest features.
Run ntk cfg set zabbix.server_version <version>
followed by ntk sys update
and agent will be automatically updated and restarted.
Specific values to monitor
Key | Type | Description |
---|---|---|
ntk.cluster.leader | (string) | Reports leader node for NtkAA cluster in this datacenter (domain) |
ntk.cluster.peercount | (int) | Reports number of available nodes in NtkAA cluster in this datacenter (domain) |
ntk.nas.v2.liccount | (int) | Reports issued licence count in NtkAS |
ntk.nas.v2.licused | (int) | Reports issued licences issued in NtkAS |
ntk.nas.thread.max | (int) | Reports preconfigured maximum number of worker threads in NtkAS |
ntk.nas.thread.running | (int) | Reports the number of running threads (spawned threads) for Puma worker of NtkAS |
ntk.nas.thread.capacity | (int) | Reports the number of requests that the server is capable of taking right now for NtkAS |
ntk.nas.thread.backlog | (int) | Reports the number of connections in that are set waiting for a free worker thread in NtkAS |
ntk.sso.fpm.conn | (int) | Total connections served since start of NtkSSO |
ntk.sso.queue.active | (int) | The number of request in the queue of pending connections |
ntk.sso.queue.max | (int) | The maximum number of requests in the queue of pending connections since FPM has started |
ntk.sso.proc.idle | (int) | The number of idle processes NtkSSO FPM |
ntk.sso.proc.active | (int) | The number of active processes NtkSSO FPM |
ntk.sso.proc.total | (int) | The number of idle + active processes of NtkSSO FPM |
ntk.sso.proc.max | (int) | The maximum number of active processes since NtkSSO has started |
ntk.sso.proc.maxreached | (int) | The number of times, the process limit has been reached, when NtkSSO pm tries to start more children |
ntk.sso.req.slow | (int) | The number of slow requests since start of NtkSSO |
ntk.cluster.live | (int) | Returns 1 if cluster container is running and quorum is reached (cluster is writable), otherwise 0 |
ntk.nas.live | (int) | Returns 1 if NAS is running and responding to requests (this is the API provider container, not management), otherwise 0 |
ntk.sso.live | (int) | Returns 1 if SSO container is running and responding to requests, 0 if there is a problem |
ntk.ap.live | (int) | Returns 1 if RADIUS proxy container is running and responding to requests, 0 if there is a problem |
ntk.rp.live | (int) | Returns 1 if reverse proxy container is running and responding to requests, otherwise 0 |
Additionaly standard alpine linux zabbix agent keys are supported. Typical things to monitor include:
vfs.fs.size[/,pfree] (float) - Free disk percentage
vm.memory.size[pused] (float) - Free RAM memory, critical as there is no swap on NtkAA
system.cpu.load[,avg5] (float) - General CPU load
Disk IO load - https://share.zabbix.com/storage-devices/linux-disk-performance-monitoring or similar
Monitoring services
Monitor all services trigger expression
{
:ntk.nas.live.last()}<1 or { :ntk.sso.live.last()}<1 or { :ntk.cluster.live.last()}<1 or { :ntk.ap.live.last()}<1 or { :ntk.rp.live.last()}<1
Not all installations require running of all services, so adjust to your setup.
Command reference
ntk zbx restart
Usage:
ntk zbx restart
Description:
Stops and starts Zabbix agent. Use this command to apply configuration changes.
ntk zbx start
Usage:
ntk zbx start [--force]
Description:
Starts Zabbix agent. Use –force option to skip configuration validation.
ntk zbx stop
Usage:
ntk zbx stop
Description:
Stops Zabbix agent. Service will be stopped and will not be started until ntk zbx start
is called. The
stopped state persists across appliance restarts.
ntk zbx validate
Usage:
ntk zbx validate
Description:
Validates Zabbix agent configuration.
ntk zbx status
Usage:
ntk zbx status
Description:
Shows status of Zabbix agent.
Log Management
Authentication Server Logs
Currently diagnostic information can be retrieved with the docker logs
command.
# Diagnostics information for the Authentication Server
$ docker logs dashboard
# Diagnostics information for the Authentication Server Mangement
$ docker logs dashboard_mgmt
SSO Server Logs
# Diagnostics information for the SSO Server
$ docker logs sso
Cluster Logs
$ docker logs consul
VRRP Logs
$ docker logs vrrp
Authentication Proxy Logs
$ docker logs auth-proxy
Audit Logs
When authentication server is running, you can configure audit events to be logged to a remote syslog via the web interface.
Submitting support requests
Use ntk sys supreq
to create a single archive containing all diagnostic information about specific cluster node. Use included scp utility to copy this archive to remote server or use scp client on your computer to retrieve archive from cluster node. Please attach this archive to support request ticket.
FAQ
How to update manually installed HTTP TLS certificate files?
To update TLS certificates follow these steps:
Use scp or similar utility to transfer certificate and key files to appliance.
Use
ntk cfg setc :ssl.pem
andntk cfg setc :ssl.key
to load data from files into appliance configuration as shown below:
ntk cfg setc :ssl.pem /home/ntkadmin/ssl.pem --load-file >/dev/null
ntk cfg setc :ssl.key /home/ntkadmin/ssl.key --load-file >/dev/null
- Restart reverse proxy service with
ntk rp restart
How to monitor appliance with SNMP?
Log in as rancher user and run the following commands to create snmp system service.
# Community in this container will be "secret"
# Generate configuration for snmp service
cat > snmp.yaml << EOF
rancher:
services:
snmp:
image: digiwhite/snmpd
restart: always
labels:
io.rancher.os.scope: system
ports:
- 161:161/udp
privileged: true
net: host
volumes:
- /proc:/host_proc
read_only: true
EOF
# Provision cloud-config
cat snmp.yaml | sudo ros config merge
# Start service
sudo ros service up snmp
# When service is no longer required, remove service with
sudo ros service rm snmp