Hardware requirements
Anaconda’s hardware recommendations ensure a reliable and performant Kubernetes cluster. The following are minimum specifications for the master and worker nodes, as well as the entire cluster.Master node | Minimum |
---|---|
CPU | 16 cores |
RAM | 64GB |
Disk space in /opt/anaconda | 500GB |
Disk space in /var/lib/gravity | 300GB |
Disk space in /tmp or $TMPDIR | 50GB |
- Disk space reserved for
/var/lib/gravity
is utilized as additional space to accommodate upgrades. Anaconda recommends having this available during installation. - The
/var/lib/gravity
volume must be mounted on local storage. Core components of Kubernetes run from this directory, some of which are extremely intolerant of disk latency. Therefore, Network-Attached Storage (NAS) and Storage Area Network (SAN) solutions are not supported for this volume. - Disk space reserved for
/opt/anaconda
is utilized for project and package storage (including mirrored packages). - Anaconda recommends that you set up the
/opt/anaconda
and/var/lib/gravity
partitions using Logical Volume Management (LVM) to provide the flexibility needed to accommodate easier future expansion. - Currently
/opt
and/opt/anaconda
must be anext4
orxfs
filesystem, and cannot be an NFS mountpoint. Subdirectories of/opt/anaconda
may be mounted through NFS. For more information, see Mounting an external file share.
Installations of Workbench that utilize an
xfs
filesystem must support d_type
file labeling to work properly. To support d-type
file labeling, set ftype=1
by running the following command prior to installing Workbench.This command will erase all data on the specified device! Make sure you are targeting the correct device and that you have backed up any important data from it before proceeding.Worker node | Minimum |
---|---|
CPU | 16 cores |
RAM | 64GB |
Disk space in /var/lib/gravity | 300GB |
Disk space in /tmp or $TMPDIR | 50GB |
When installing Workbench on a system with multiple nodes, verify that the clock of each node is in sync with the others prior to installation. Anaconda recommends using the Network Time Protocol (NTP) to synchronize computer system clocks automatically over a network. For step-by-step instructions, see How to Synchronize Time with Chrony NTP in Linux.
Disk IOPS requirements
Master and worker nodes require a minimum of 3000 concurrent Input/Output operations Per Second (IOPS).Hard disk manufacturers report sequential IOPS, which are different than concurrent IOPS. On-premise installations require servers with disks that support a minimum of 50 sequential IOPS. Anaconda recommends using Solid State Drive (SSD) or better.
Cloud performance requirements
Requirements for running Workbench in the cloud relate to compute power and disk performance. Make sure your chosen cloud platform meets these minimum specifications:Anaconda recommends an instance type no smaller than
m4.4xlarge
for both master and worker nodes. You must have a minimum of 3000 IOPS.Operating system requirements
Workbench currently supports the following Linux versions:- RHEL/CentOS 7.x, 8.x
- Ubuntu 16.04
- SUSE 12 SP2, 12 SP3, 12 SP5 (Requires you set
DefaultTasksMax=infinity
in/etc/systemd/system.conf
)
Some versions of the RHEL 8.4 AMI on AWS are bugged due to a combination of a bad
ip rule
and the networkmanager service. Remove the bad rule and disable the networkmanager service prior to installation.Security requirements
- If your Linux system utilizes an antivirus scanner, make sure the scanner excludes the
/var/lib/gravity
volume from its security scans. - Installation requires that you have
sudo
access. - Nodes running CentOS or RHEL must make sure that Security Enhanced Linux (SELinux) set to either
disabled
orpermissive
mode in the/etc/selinux/config
file.
Check the status of SELinux by running the following command:
Configuring SELinux
Configuring SELinux
- Open the
/etc/selinux/config
file using your preferred file editor. - Find the the line that starts with
SELINUX=
and set it to eitherdisabled
orpermissive
. - Save and close the file.
- Reboot your system for changes to take effect.
Kernel module requirements
Kubernetes relies on certain functionalities provided by the Linux kernel. The Workbench installer verifies that the following kernel modules (required for Kubernetes to function properly) are present, and notifies you if any are not loaded.Linux Distribution | Version | Required Modules |
---|---|---|
CentOS | 7.2 | bridge, ebtable_filter, ebtables, iptable_filter, iptable_nat, overlay |
CentOS | 7.3-7.7, 8.0 | br_netfilter, ebtable_filter, ebtables, iptable_filter, iptable_nat, overlay |
RedHat Linux | 7.2 | bridge, ebtable_filter, ebtables, iptable_filter, iptable_nat |
RedHat Linux | 7.3-7.7, 8.0 | br_netfilter, ebtable_filter, ebtables, iptable_filter, iptable_nat, overlay |
Ubuntu | 16.04 | br_netfilter, ebtable_filter, ebtables, iptable_filter, iptable_nat, overlay |
Suse | 12 SP2, 12 SP3, 12 SP5 | br_netfilter, ebtable_filter, ebtables, iptable_filter, iptable_nat, overlay |
Module Name | Purpose |
---|---|
bridge | Enables Kubernetes iptables-based proxy to operate |
br_nerfilter | Enables Kubernetes iptables-based proxy to operate |
overlay | Enables the use of the overlay or overlay2 Docker storage driver |
ebtable_filter | Allows a service to communicate back to itself via internal load balancing when necessary |
ebtables | Allows a service to communicate back to itself via internal load balancing when necessary |
iptable_filter | Ensures the firewall rules set up by Kubernetes function properly |
iptable_nat | Ensures the firewall rules set up by Kubernetes function properly |
Verify a module is loaded by running the following command:If the command produces a return, the module is loaded.If necessary, run the the following command to load a module:
If your system does not load modules at boot, you must run the following command—for each module—to ensure they are loaded on every reboot:
System control settings
Workbench requires the following Linuxsysctl
settings to function properly:
sysctl setting | Purpose |
---|---|
net.bridge.bridge-nf-call-iptables | Communicates with bridge kernel module to ensure Kubernetes iptables-based proxy operates |
net.bridge.bridge-nf-call-ip6tables | Communicates with bridge kernel module to ensure Kubernetes iptables-based proxy operates |
fs.may_detach_mounts | Allows the unmount operation to complete even if there are active references to the filesystem remaining |
net.ipv4.ip_forward | Required for internal load balancing between servers to work properly |
fs.inotify.max_user_watches | Set to 1048576 to improve cluster longevity |
If necessary, run the following command to enable a system control setting:To persist system settings on boot, run the following for each setting:
GPU requirements
Workbench requires that you install a supported version of the NVIDIA Compute Unified Device Architecture (CUDA) driver on the host OS of any GPU worked node. Currently, Workbench supports the following CUDA driver versions:- CUDA
10.2
- CUDA
11.2
- CUDA
11.4
- CUDA
11.6
Notify your Anaconda Implementation team member which CUDA version you intend to use, so they can provide the correct installer.
- Use the package manager or the Nvidia runfile to download the file directly.
- For SLES, CentOS, and RHEL, you can get a supported driver using
rpm (local)
orrpm (network)
. - For Ubunutu, you can get a driver using
deb (local)
ordeb (network)
.
- Tesla V100 (recommended)
- Tesla P100 (adequate)
11.6
:
- A-Series: NVIDIA A100, NVIDIA A40, NVIDIA A30, NVIDIA A10
- RTX-Series: RTX 8000, RTX 6000, NVIDIA RTX A6000, NVIDIA RTX A5000, NVIDIA RTX A4000, NVIDIA T1000, NVIDIA T600, NVIDIA T400
- HGX-Series: HGX A100, HGX-2
- T-Series: Tesla T4
- P-Series: Tesla P40, Tesla P6, Tesla P4
- K-Series: Tesla K80, Tesla K520, Tesla K40c, Tesla K40m, Tesla K40s, Tesla K40st, Tesla K40t, Tesla K20Xm, Tesla K20m, Tesla K20s, Tesla K20c, Tesla K10, Tesla K8
- M-Class: M60, M40 24GB, M40, M6, M4
Network requirements
Workbench requires the following network ports to be externally accessible:External ports
External ports
Port | Protocol | Description |
---|---|---|
80 | TCP | Workbench UI (plaintext) |
443 | TCP | Workbench UI (encrypted) |
32009 | TCP | Operations Center Admin UI |
Install ports
Install ports
Port | Protocol | Description |
---|---|---|
4242 | TCP | Bandwidth checker utility |
61009 | TCP | Install wizard UI access required during cluster installation |
61008, 61010, 61022-61024 | TCP | Installer agent ports |
Cluster communication ports
Cluster communication ports
Port | Protocol | Description |
---|---|---|
53 | TCP and UDP | Internal cluster DNS |
2379, 2380, 4001, 7001 | TCP | Etcd server communication |
3008-3012 | TCP | Internal Workbench service |
3022-3025 | TCP | Teleport internal SSH control panel |
3080 | TCP | Teleport Web UI |
5000 | TCP | Docker registry |
6443 | TCP | Kubernetes API Server |
6990 | TCP | Internal Workbench service |
7496, 7373 | TCP | Peer-to-peer health check |
7575 | TCP | Cluster status gRPC API |
8081, 8086-8091, 8095 | TCP | Internal Workbench service |
8472 | UDP | Overlay network |
9080, 9090, 9091 | TCP | Internal Workbench service |
10248-10250, 10255 | TCP | Kubernetes components |
30000-32767 | TCP | Kubernetes internal services range |
There are various tools you can use to configure firewalls and open required ports, including
iptables
, firewall-cmd
, susefirewall2
, and more!10.244.0.0/16
pod subnet and 10.100.0.0/16
service subnet are accessible to every node in the cluster, and grant all nodes the ability to communicate via their primary interface.
For example, if you’re using iptables
:
- repo.anaconda.com
- anaconda.org
- conda.anaconda.org
- binstar-cio-packages-prod.s3.amazonaws.com
- https://repo.anaconda.com (or for older versions of Navigator and conda)
- https://conda.anaconda.org if any users will use conda-forge and other channels on Anaconda.org
- google-public-dns-a.google.com (8.8.8.8:53) to check internet connectivity with Google Public DNS
TLS/SSL certificate requirements
Workbench uses certificates to provide transport layer security for the cluster. Self-signed certificates are generated during the initial installation. Once installation is complete, you can configure the platform to use your organizational TLS/SSL certificates. You can purchase certificates commercially, or generate them using your organization’s internal public key infrastructure (PKI) system. When using an internal PKI-signed setup, the CA certificate is inserted into the Kubernetes secret. In either case, the configuration will include the following:- A certificate for the root certificate authority (CA)
- An intermediate certificate chain
- A server certificate
- A certificate private key
DNS requirements
Workbench assigns unique URL addresses to deployments by combining a dynamically generated universally unique identifier (UUID) with your organization’s domain name, like this:https://uuid001.anaconda.yourdomain.com
.
This requires the use of wildcard DNS entries that apply to a set of domain names such as *.anaconda.yourdomain.com
.
For example, if you are using the domain name anaconda.yourdomain.com
with a master node IP address of 12.34.56.78
, the DNS entries would be as follows:
The wildcard subdomain’s DNS entry points to the Workbench master node.
/etc/hosts
entries to the gravity environment.
If If necessary, stop and disable
dnsmasq
is installed on the master node or any worker nodes, you’ll need to remove it from all nodes prior to installing Workbench.Verify dnsmasq
is disabled by running the following command:dnsmasq
, run the following commands:Browser requirements
Workbench supports the following web browsers:- Chrome 39+
- Firefox 49+
- Safari 10+
Verifying system requirements
The installer performs pre-installation checks, and only allows installation to continue on nodes that are configured correctly, and include the required kernel modules. If you want to perform the system check yourself prior to installation, you can run the following commands from the installer directory,~/anaconda-enterprise-<VERSION>
, on your intended master and worker nodes:
To perform system checks on the master
node, run the following command as sudo or
root user:
worker
node, run the following command as sudo or
root user:
Pre-installation checklist
Anaconda has created this pre-installation checklist to help you verify that you have properly prepared your environment prior to installation. You can run the system verification checks to automatically verify many of the requirements for you.System verficication checks are not comprehensive, so make sure you manually verify the remaining requirements.
Gravity pre-inallation checklist
Gravity pre-inallation checklist
- All nodes in the cluster meet the minimum or recommended specifications for CPU, RAM, and disk space.
- All nodes in the cluster meet the minimum IOPS required for reliable performance.
- All cluster nodes are operating the same version of the OS, and that the OS version is supported by Workbench.
- NTP is being used to synchronize computer system clocks, and all nodes are in sync.
- The user account performing the installation has
sudo
access on all nodes and is not a root user. - All required kernel modules are loaded.
- The sysctl settings are configured correctly.
- Any GPUs to be used with Workbench have a supported NVIDIA CUDA driver installed.
- The system meets all network port requirements, whether the specified ports need to be open internally, externally, or during installation only.
- The firewall is configured correctly, and an rules designed to limit traffic have been temporarily disabled until Workbench is installed and verified.
- If necessary, the domains required for online package mirroring have been allowlisted.
- The final TLS/SSL certificates <grav_tls_ssl_reqs> to be installed with Workbench have been obtained, including the private keys.
- The Workbench
A
orCNAME
domain record is fully operational, and points to the IP address of the master node. - The wildcard DNS entry for Workbench is also fully operational, and points to the IP address of the master node. More information about the wildcard DNS requirements can be found here.
- The
/etc/resolv.conf
file on all the nodes does not include therotate
option. - Any existing installations of Docker (and
dockerd
),dnsmasq
, andlxd
have been removed from all nodes, as they will conflict with Workbench. - All web browsers to be used to access Workbench are supported by the platform.