Before you can use the Tanzu CLI or installer interface to deploy a management cluster, you must prepare your vSphere environment. You must make sure that vSphere meets the general requirements, and import the base image templates from which Tanzu Kubernetes Grid creates cluster node VMs. Each base image template contains a version of a machine OS and a version of Kubernetes.
ImportantFor TKG deployments to vSphere, VMware recommends that you use the vSphere with Tanzu Supervisor. Using TKG with a standalone management cluster is only recommended for the use cases listed in When to Use a Standalone Management Cluster in About TKG.
kubectl
installed. See Install the Tanzu CLI and Kubernetes CLI for Use with Standalone Management Clusters.
tanzu
, kubectl
and other commands.vsphere-zones.yaml
file that defines the zones’ FailureDomain
and DeploymentZone
objects.
FailureDomain
and DeploymentZone
Objects in Kubernetes under Running Clusters Across Multiple Availability Zones.A vSphere network* with:
VSPHERE_CONTROL_PLANE_ENDPOINT
, or if you are using NSX Advanced Load Balancer for your control plane endpoint, let the address be set automatically from an address pool.CLUSTER_API_SERVER_PORT
or for environments with NSX Advanced Load Balancer, VSPHERE_CONTROL_PLANE_ENDPOINT_PORT
variable when deploying the cluster.~/.config/tanzu/tkg/bom/
and its name includes the Tanzu Kubernetes Grid version, for example tkg-bom-v2.4.1+vmware.1.yaml
for v2.4.1.date
command to see the timezone settings.esxcli system time set
.NSX Advanced Load Balancer (ALB) installed in your vSphere instance, if you want to use NSX ALB as the load balancer and the endpoint provider for control plane HA. See Install NSX Advanced Load Balancer.
If your vSphere environment runs VMware NSX, you can use the NSX interfaces when you deploy management clusters. Make sure that your NSX setup includes a segment on which DHCP is enabled. Make sure that NTP is configured on all ESXi hosts, on vCenter Server, and on the bootstrap machine.
*Or see Prepare an Internet-Restricted Environment for installing without external network access.
The table below describes sizing examples for management clusters on vSphere. Use this data as guidance to ensure your management cluster is scaled to handle the number of workload clusters that you plan to deploy. The Workload cluster VM size column lists the VM sizes that were used for the examples in the Can manage… column.
Management cluster plan | Management cluster VM size | Can manage… | Workload cluster VM size |
---|---|---|---|
3 control plane nodes and 3 worker nodes | Control plane nodes:
Worker nodes:
|
Examples:
|
Control plane nodes:
Worker nodes:
|
3 control plane nodes and 3 worker nodes | Control plane nodes:
Worker nodes:
|
Example: One workload cluster, deployed with 3 control plane and 500 worker nodes | Control plane nodes:
Worker nodes:
|
3 control plane nodes and 3 worker nodes | Control plane nodes:
Worker nodes:
|
Example: 200 workload clusters, each cluster deployed with 3 control plane and 5 worker nodes | Control plane nodes:
Worker nodes:
|
Also, see Minimum VM Sizes for Cluster Nodes below.
On vSphere 8, the vSphere with Tanzu feature includes a Supervisor that you can use as a management cluster for Tanzu Kubernetes Grid. This means that on vSphere 8, you do not need to use tanzu management-cluster create
or tanzu mc create
to deploy a management cluster if vSphere with Tanzu and the Supervisor are enabled. Deploying a Tanzu Kubernetes Grid management cluster to vSphere 8 when vSphere with Tanzu is not enabled is supported, but the preferred option is to enable vSphere with Tanzu and use the built-in Supervisor Cluster if possible. The vSphere with Tanzu Supervisor is closely integrated with vSphere, so offers a more streamlined user experience than using a standalone management cluster. However, using a standalone management cluster on vSphere offers more configuration and customization options than a Supervisor.
ImportantThe versions of the Tanzu CLI that are compatible with TKG 2.x and with the vSphere with Tanzu Supervisor in vSphere 8 are not compatible with the Supervisor Cluster in vSphere 7. To use the Tanzu CLI with a vSphere with Tanzu Supervisor Cluster on vSphere 7, use the Tanzu CLI version from TKG v1.6. To use the versions of the Tanzu CLI that are compatible with TKG 2.x with Supervisor, upgrade to vSphere 8. You can deploy a standalone TKG 2.x management cluster to vSphere 7 if a vSphere with Tanzu Supervisor Cluster is not present. For information about compatibility between the Tanzu CLI and VMware products, see the Tanzu CLI Documentation.
The Tanzu CLI works with both management clusters provided by vSphere with Tanzu and with standalone management clusters deployed by Tanzu Kubernetes Grid on Azure, Amazon Web Services (AWS), and vSphere when vSphere with Tanzu is not enabled, letting you deploy and manage workload clusters across multiple infrastructures using a single tool. For information about how to use the Tanzu CLI with a Supervisor, see Connect the Tanzu CLI to the Supervisor in Creating and Managing TKG 2.4 Workload Clusters with the Tanzu CLI.
For information about the vSphere with Tanzu feature in vSphere 8, see the vSphere with Tanzu 8.0 documentation.
NoteOn Azure VMware Solution, you cannot create a Supervisor Cluster, and need to deploy a management cluster to run
tanzu
commands.
Each management cluster and workload cluster that you deploy to vSphere requires one static virtual IP address for external requests to the cluster’s API server. You must be able to assign this IP address, so it cannot be within your DHCP range, but it must be in the same subnet as the DHCP range.
The cluster control plane’s Kube-Vip pod uses this static virtual IP address to serve API requests, and the API server certificate includes the address to enable secure TLS communication. In workload clusters, Kube-Vip runs in a basic, Layer-2 failover mode, assigning the virtual IP address to one control plane node at a time. In this mode, Kube-Vip does not function as a true load balancer for control plane traffic.
Tanzu Kubernetes Grid can use Kube-Vip as a load balancer for workloads in workload clusters (Technical Preview). You cannot use Kube-VIP as a LoadBalancer
service on Windows-based clusters. For more information, see Kube-VIP Load Balancer.
To load-balance workloads on vSphere, use NSX Advanced Load Balancer, also known as Avi Load Balancer, Essentials Edition.
ImportantOn vSphere 8, to use NSX Advanced Load Balancer with a TKG standalone management cluster and its workload clusters you need NSX ALB v22.1.2 or later and TKG v2.1.1 or later.
Before you can deploy a cluster to vSphere, you must import into vSphere a base image template containing the OS and Kubernetes versions that the cluster nodes run on. For each supported pair of OS and Kubernetes versions, VMware publishes a base image template in OVA format, for deploying clusters to vSphere. After you import the OVA into vSphere, you must convert the resulting VM into a VM template.
Supported base images for cluster nodes depend on the type of cluster, as follows:
Management Cluster: OVA must have Kubernetes v1.27.5, the default version for Tanzu Kubernetes Grid v2.4.1. So it must be one of the following:
Ubuntu v20.04 Kubernetes v1.27.5 OVA
NoteIn Tanzu Kubernetes Grid v2.4.1, the Ubuntu OVA image uses the Unified Extensible Firmware Interface (UEFI) booting mode.
Photon v3 Kubernetes v1.27.5 OVA
A custom OVA with a custom Tanzu Kubernetes release (TKr), as described in Build Machine Images.
To import a base image template into vSphere:
Download a Tanzu Kubernetes Grid OVA for the cluster nodes. For the management cluster, this must be one of the Kubernetes v1.27.5 OVA downloads.
ImportantMake sure you download the most recent OVA base image templates in the event of security patch releases.
You can find updated base image templates that include security patches on the Tanzu Kubernetes Grid product download page.
In the vSphere Client, right-click an object in the vCenter Server inventory, select Deploy OVF template.
Follow the installer prompts to deploy a VM from the OVA.
NoteIf you select thick provisioning as the disk format, when Tanzu Kubernetes Grid creates cluster node VMs from the template, the full size of each node’s disk will be reserved. This can rapidly consume storage if you deploy many clusters or clusters with many nodes. However, if you select thin provisioning, as you deploy clusters this can give a false impression of the amount of storage that is available. If you select thin provisioning, there might be enough storage available at the time that you deploy clusters, but storage might run out as the clusters run and accumulate data.
When the OVA deployment finishes, right-click the VM and select Template > Convert to Template.
ImportantDo not power on the VM before you convert it to a template.
In the VMs and Templates view, right-click the new template, select Add Permission, and assign the tkg-user
to the template with the TKG
role.
For information about how to create the user and role for Tanzu Kubernetes Grid, see Required Permissions for the vSphere Account below.
Repeat the procedure for each of the Kubernetes versions for which you downloaded the OVA file.
The vCenter Single Sign On account that you provide to Tanzu Kubernetes Grid when you deploy a management cluster must have the correct permissions in order to perform the required operations in vSphere.
It is not recommended to provide a vSphere administrator account to Tanzu Kubernetes Grid, because this provides Tanzu Kubernetes Grid with far greater permissions than it needs. The best way to assign permissions to Tanzu Kubernetes Grid is to create a role and a user account, and then to grant that user account that role on vSphere objects.
NoteIf you are deploying workload clusters to vSphere 7 or 8 and vSphere with Tanzu is enabled, you must set the Global > Cloud Admin permission in addition to the permissions listed below. If you intend to use Velero to back up and restore workload clusters, you must also set the permissions listed in Credentials and Privileges for VMDK Access in the Virtual Disk Development Kit Programming Guide.
In the vSphere Client, go to Administration > Access Control > Roles, and create a new role, for example TKG
, with the following permissions.
vSphere Object | Required Permission |
---|---|
Cns | Searchable |
Datastore | Allocate space Browse datastore Low level file operations |
Global (if using Velero for backup and restore) | Disable methods Enable methods Licenses |
Network | Assign network |
Profile-driven storage | Profile-driven storage view |
Resource | Assign virtual machine to resource pool |
Sessions | Message Validate session |
Virtual machine | Change Configuration > Add existing disk Change Configuration > Add new disk Change Configuration > Add or remove device Change Configuration > Advanced configuration Change Configuration > Change CPU count Change Configuration > Change Memory Change Configuration > Change Settings Change Configuration > Configure Raw device Change Configuration > Extend virtual disk Change Configuration > Modify device settings Change Configuration > Remove disk Change Configuration > Toggle disk change tracking* Edit Inventory > Create from existing Edit Inventory > Remove Interaction > Power On Interaction > Power Off Provisioning > Allow read-only disk access* Provisioning > Allow virtual machine download* Provisioning > Deploy template Snapshot Management > Create snapshot* Snapshot Management > Remove snapshot* *Required to enable the Velero plugin, as described in Back Up and Restore Management and Workload Cluster Infrastructure. You can add these permissions when needed later. |
vApp | Import |
In Administration > Single Sign On > Users and Groups, create a new user account in the appropriate domain, for example tkg-user
.
In the Hosts and Clusters, VMs and Templates, Storage, and Networking views, right-click the objects that your Tanzu Kubernetes Grid deployment will use, select Add Permission, and assign the tkg-user
with the TKG
role to each object.
Configure the sizes of your management and workload cluster nodes depending on cluster complexity and expected demand. You can set them to small
, medium
, large
, or extra-large
as defined in Predefined Node Sizes.
For all clusters on vSphere, you configure these with the SIZE
, CONTROLPLANE_SIZE
, and WORKER_SIZE
cluster configuration variables. Or for greater granularity, you can use the VSPHERE_*
_DISK_GIB
, _NUM_CPUS
, and _MEM_MIB
configuration variables.
For management clusters, the installer interface Instance Type field also configures node VM sizes.
For single-worker management and workload clusters running sample applications, use the following minimum VM sizes:
small
medium
In order for the Tanzu CLI to connect to vSphere from the machine on which you run it, you must provide the public key part of an SSH key pair to Tanzu Kubernetes Grid when you deploy the management cluster. If you do not already have one on the machine on which you run the CLI, you can use a tool such as ssh-keygen
to generate a key pair.
On the machine on which you will run the Tanzu CLI, run the following ssh-keygen
command.
ssh-keygen -t rsa -b 4096 -C "email@example.com"
At the prompt Enter file in which to save the key (/root/.ssh/id_rsa):
press Enter to accept the default.
Add the private key to the SSH agent running on your machine, and enter the password you created in the previous step.
ssh-add ~/.ssh/id_rsa
Open the file .ssh/id_rsa.pub
in a text editor so that you can easily copy and paste it when you deploy a management cluster.
If your vSphere environment uses untrusted, self-signed certificates to authenticate connections, you must verify the thumbprint of the vCenter Server when you deploy a management cluster. If your vSphere environment uses trusted certificates that are signed by a known Certificate Authority (CA), you do not need to verify the thumbprint.
You can use your Web browser’s certificate viewer to obtain the vSphere certificate thumbprint.
Access the certificate viewer by clicking on the Secure (padlock) icon to the left of the Web address in the URL field.
The next steps depend on which browser you use. For example, in Google Chrome, you select Connection is secure > Certificate is valid to see the certificate details, including the thumbprint.
Record the SHA-1 Fingerprint value from the browser. If it contains spaces between each hex pair, substitute a :
character for each space, for example 6D:4A:DC:6C:C4:43:73:BB:DF:9A:32:68:67:56:F9:96:02:08:64:F4
.
You can use this thumbprint string to verify it when you deploy a management cluster from the installer interface, or provide it to the VSPHERE_TLS_THUMBPRINT
option when you deploy clusters from a configuration file.
To deploy a management cluster that supports IPv6 in an IPv6 networking environment:
Configure Linux to accept router advertisements to ensure the default IPv6 route is not removed from the routing table when the Docker service starts. For more information, see Docker CE deletes IPv6 Default route. sudo sysctl net.ipv6.conf.eth0.accept_ra=2
Create a masquerade rule for bootstrap cluster to send outgoing traffic from the bootstrap cluster: sudo ip6tables -t nat -A POSTROUTING -s fc00:f853:ccd:e793::/64 ! -o docker0 -j MASQUERADE
For more information about masquerade rules, See MASQUERADE.
Deploy the management cluster by running tanzu mc create
, as described in Deploy Management Clusters from a Configuration File.
TKG_IP_FAMILY
and other variables as described in Configure for IPv6.To deploy a standalone management and workload clusters to run across multiple availability zones (AZs) in vSphere, you need to:
Create or identify either of the following sets of objects in vSphere:
Tag the objects to associate them with a region and its AZs in Kubernetes, as described in Prepare Regions and AZs in vSphere.
For production deployments, it is strongly recommended to enable identity management for your clusters:
If you are using Tanzu Kubernetes Grid in an environment with an external internet connection, once you have set up identity management, you are ready to deploy management clusters to vSphere.