In this blog I’ll explain how to install Rancher on a high-available k3s kubernetes cluster with an embedded etcd database and kube-vip as the load balancer in front of the kubernetes controlplane.

For those who don’t know Rancher, Rancher is an enterprise container management platform built for organizations with multiple teams that deploy containers in production across multiple cloud environments. Rancher makes it easy to run Kubernetes everywhere, meet IT requirements and empower DevOps teams. It’s a great opensource alternative for VMware vSphere with Tanzu, VMware Tanzu Kubernetes Grid with Tanzu Mission Control or Redhat Openshift.

Rancher provides an intuitive user interface to manage application workloads. The user does not need to have in-depth knowledge of Kubernetes concepts to start using Rancher. Rancher catalog contains a set of useful DevOps tools. Rancher is certified with a wide selection of cloud native ecosystem products, including, for example, security tools, monitoring systems, container registries, and storage and networking drivers.

The following figure illustrates the role Rancher plays in IT and DevOps organizations. Each team deploys their applications on the public or private clouds they choose. IT administrators gain visibility and enforce policies across all users, clusters, and clouds.

Install RancherLet’s get started!

k3s prerequisites

To install a k3s cluster with a high-available control plane, you need a minimum of three nodes. In my case, I deployed three Virtual Machines with a fixed IP address on VMware vSphere 7.0u1 with Ubuntu 20.04 installed and a user vmware configured.

You also need to create four DNS entries in your domain. Three for the k3s nodes and one for the kube-vip load balancer VIP address. For example: – (VIP) – – –

Also make sure you can login to the k3s nodes remotely using a username and ssh key. To realize that, first create or use an ssh key and install this on the remote machine.

#On your local machine

git clone git://

cd generate-and-send-ssh-key

chmod +x

./ --file ~/.ssh/sshkey --user vmware --host

Login to the first node and configure SSH to use Public Key Authentication.

ssh vmware@

sudo sed -i "s/.*PubkeyAuthentication.*/PubkeyAuthentication yes/g" /etc/ssh/sshd_config

sudo systemctl restart sshd

Add user vmware to the sudo group and change the sudoers file.

sudo adduser vmware sudo

sudo visudo
#replace this line

#by this line

Disable swap. Then logout and reboot.

sudo swapoff -a

sudo sed -i '/ swap / s/^\(.*\)$/#\1/g' /etc/fstab

sudo reboot

Repeat these steps on the second and third node.

TIP! Use Hasicorp Packer to create vSphere templates with above settings (and others) already in place. Then you only have to deploy a VM from this template and apply a fixed IP address making use of a vSphere customization specification. This is what I did ;-)

Verify you can login to the first node using ssh.

install Rancher

Install k3s using k3sup

For the next steps, make sure you have installed the following tools on your local machine.

K3s is a fully CNCF compliant lightweight Kubernetes distribution. Easy to install, half the memory, all in a binary of less than 100 MB.

k3sup is a light-weight utility to get from zero to KUBECONFIG with k3s on any local or remote VM. All you need is ssh access and the k3sup binary to get kubectl access immediately.

Install the first k3s master server using k3sup.

k3sup install --ip --tls-san --cluster --k3s-channel latest --merge --local-path $HOME/.kube/config --context=k3s-ha-cluster --user vmware --ssh-key $HOME/.ssh/sshkey
  • --tls-san is required to advertise the kube-vip VIP address, so that K3s will create a valid certificate for the API server.
  • the --k3s-channel is specifying the latest version of K3s, which in this instance will be 1.20, by the time you run this tutorial, it may have changed, in which can you can give 1.20 as the channel, or a specific version with --k3s-version
  • note the --cluster flag, which tells the server to use etcd to create a cluster for the servers we will join later on
  • --local-path, --context and --merge all allow us to merge the KUBECONFIG from the K3s to our local file

install Rancher

Check if k3s kubernetes is installed succesfully.

kubectx k3s-ha-cluster

kubectl get nodes

install Rancher

Installing kube-vip

Kube-vip is a lightweight solution that provides Kubernetes Virtual IP and Load-Balancer for both control plane and Kubernetes services.

Login as root into the first k3s node and apply RBAC settings for kube-vip.

curl -s > /var/lib/rancher/k3s/server/manifests/kube-vip-rbac.yaml

By downloading this manifest and placing it in the k3s manifest directory, this manifest will automatically be applied by k3s and create a kube-vip serviceAccount, kube-vip-role clusterRole and a kube-vip-binding clusterRoleBinding.

install Rancher

Next step is to fetch the kube-vip container, create an kube-vip alias and generate a kube-vip manifest which will deploy a daemonset.

ctr image pull

alias kube-vip="ctr run --rm --net-host vip /kube-vip"

export VIP=

export INTERFACE=ens192

kube-vip manifest daemonset --arp --interface $INTERFACE --address $VIP --controlplane --leaderElection --taint --inCluster | sudo tee /var/lib/rancher/k3s/server/manifests/kube-vip.yaml

#Edit the kube-vip.yaml
      - effect: NoSchedule
        operator: Exists

install Rancher install Rancher install Rancher install Rancher

Check if kube-vip is correctly installed and ping the VIP address.

install Rancher

Logout and change your local KUBECONFIG to use the VIP address.

Joining the other nodes

Next step is to join the second and third node to the k3s cluster using k3sup.

k3sup join --ip --server --server-ip --k3s-channel latest --user vmware --ssh-key $HOME/.ssh/sshkey

k3sup join --ip --server --server-ip --k3s-channel latest --user vmware --ssh-key $HOME/.ssh/sshkey

Check if the nodes are added to the cluster as additional master nodes.

install Rancher

Now you have a high-available k3s cluster with an embedded etcd database using kube-vip as load balancer in front of the kubernetes control plane.

Installing Rancher

With the high-available kubernetes cluster in place it’s finally time to install Rancher.

Make sure you have installed helm on your local machine!

Let’s start with some prerequisites such as creating some namespaces and adding helm repos for Rancher and cert-manager.

kubectl create namespace cattle-system

kubectl create namespace cert-manager

kubectl apply --validate=false -f

helm repo add rancher-latest

helm repo add jetstack

helm repo update

install Rancher install Rancher

Next step is to install cert-manager using helm.

helm install cert-manager jetstack/cert-manager --namespace cert-manager --version v1.0.4

kubectl get pods --namespace cert-manager

install Rancher

The final step is to install Rancher using helm.

helm install rancher rancher-latest/rancher --namespace cattle-system --set

kubectl -n cattle-system rollout status deploy/rancher

kubectl -n cattle-system get deploy rancher

install Rancher

Now try to login to the GUI of Rancher by opening a web browser and pointing it at the DNS entry of your kube-vip VIP address. In my case, (

install Rancher

If this page pops up then Rancher is installed correctly.

Enter a password for the admin account. Choose the default view “I want to create or manage multiple clusters”. Select the “I agree to the terms and conditions for using Rancher” and click Continue.

install Rancher

Enter the Rancher Server URL and click Save URL.

install Rancher

And there you have it, Rancher is up and running, ready to deploy and manage kubernetes clusters and their workloads in the private and/or public cloud of your choice.

Happy containering!