Typically Minikube is a utility that most developers use when they are learning Kubernetese. Though it’s a great tool, Minikube runs a single-node Kubernetes cluster inside a VM. Ideally, Minikube is not suited when you want to learn and execute multi-cluster Use Case scenarios. This motivated me to build and configure a multi-cluster Kubernetes cluster in my local VM using a Virtual box with which I have full control.
Although you can find many tools, websites, and practice LABS where you can learn about Kubernetes, having your own cluster make it so easy and convenient compared to cloud-based tools.
In this tutorial, I will help you install 2 nodes Kubernetes cluster on your own computer so you don’t have to sign-up for AWS or Azure account, spin up computer servers and make it available when you need it.
Following instructions will work out of the box on the latest Mac OS X however similar steps may work on any computer with virtualization software installed.
So let’s get started.
This Tutorial was originally published in Medium blog
My router network is configured at 192.168.1.0/24
You can check your router detail, by checking the Network settings in your system. In Mac, its Network → Advance → TCP/IP Tab as shown below
So this is what we are going to do. We will be creating multi-cluster VM instances of one Master and worker node. These VM instances will have the OS of Ubuntu 20.4, and each of these VM instances will have two network interfaces, one Bridge adapter, and another NAT Adapter.
With Bridge adapter, the VM will be in the same network as the router network, and it will enable me to ssh VM instance and no need to login to the console of these VM instances.
The NAT adapter, which will be assigned in a separate subnet, is how the Master and Worker node will communicate. Also, with NAT adapter, VM instances will access the external network, but no direct access to VM is possible.
The CIDR range provided to the NAT adapters ( 192.168.10.0/24) differs from the Home router Network ( 192.168.1.0/24), so there won’t be any conflict.
This architectural setup is what we are going to achieve by end of this article
Add on the CIDR range (192.168.10.0/24)as we discussed earlier
Let’s configure the VM, create a new Machine(VM instance). You can name the machine and mention the path for the folder on the host machine (Machine folder)
I have set the size as 20GB, but the default 10GB should also be fine.
Also set the CPU to 4 for master, again 2 should be fine if you are not intending for heavy workloads
Now is the time to configure Network, both Bridge and NAT adapter, so navigate to network tab set Adapter one as Bridge Network and adapter two as NAT Network( as configured earlier)
Navigate to storage and ensure that when the VM boots up, it loads the Ubuntu image which we downloaded earlier, so select the ISO image
Now you can start the virtual machine and can select the language of your choice. I am selecting English here
You will see the two network interfaces details which were specified earlier. The two IP addresses, one as part of the address pool of the Home router network(enp0s3) and another part of the NAT CIDR address pool we defined earlier (enp0s8)
Next , you can continue with default
Next section remove the option Set up this disk as LVM group
Next enter your username and password details which will be used to log into the system
Next , select install of OpenSSH server
No Need to select any feature as to keep the configuration minimal
Click on Done and the installation starts
The installation will take some time. Once the installation is done, try to SSH the system using the command (for Mac)
Note here 192.168.1.117 IP is the Ip from the pool of bridge network as configured earlier.
Its a good practice to make the IP address of instances static so for that log into VM instance ( Via SSH) , navigate to /etc/network/interfaces, you can do sudo /etc/network/interfaces and add the following, do same for worker node, and please note the ip address, the gateway will be different in your case
Now you need to start configuring the Master and worker node. We will create a cluster with kubeadm. Some commands need to run on both Master and worker node, only Master Node and Only worker node. I will tag these commands accordingly( Master and Worker / Master Only / Worker Only)
Commands to be executed ( Master+Worker Node)
sudo apt-get update
2. Install Docker
sudo apt-get install docker.io -y
3. Get Curl
sudo apt-get install apt-transport-https curl -y
4. Add Kubernetes repository for Ubuntu 20.04, Add K8S key and Repo
curl -s https://packages.cloud.google.com/apt-key.gpg | sudo apt-key add-cat << EOF | sudo tee /etc/apt/sources.list.d/kubernetes.list deb https://apt.kubernetes.io/ kubernetes-xenial main EOF
5. Install the required packages
sudo apt-get update sudo apt-get install -y kubelet=1.20.1–00 sudo apt-get install -y kubeadm=1.20.1–00 sudo apt-get install -y kubectl=1.20.1–00 sudo apt-mark hold kubelet kubeadm kubectl
I have used Kubernetes version 1.20.1, you can use the appropriate version to update the package repository and Install K8S components
6. Disable SWAP
sudo swapoff -a edit /etc/fstab to remove the swap entry
Commands to be executed ( Only Master Node)
sudo kubeadm init — control-plane-endpoint kube-master:6443 — pod-network-cidr 10.10.0.0/16
In this case, the POD CIDR is taken as 10.10.0.0/16, these IP address is completely internal to kubernetese so can assign any IP
2. Set the kubectl context auth to connect to the cluster
mkdir -p $HOME/.kube sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config sudo chown $(id -u):$(id -g) $HOME/.kube/config
3. Install Calico networking add -on , You can read more about Calico, Download the Calico networking manifest
curl https://docs.projectcalico.org/manifests/calico.yaml -O
4. Uncomment the CALICO_IPV4POOL_CIDR variable in the manifest and set it to the same value as your chosen pod CIDR(10.10.0.0/16)
5. Apply Calico configuration
kubectl apply -f calico.yaml
6. Wait for the PODS to be in running state and if you run command
kubectl get nodes
you will see only the Kube master in running state and status is ready
7. Now you need to add the worker node to the cluster, so if you go back to the init set( step1), you should see kubeadm join command which can be used to join the worker node
Commands to be executed ( Only Worker Node)
kubectl get nodes
run the following command
kubectl run nginx — image=nginx
Congratulations!!! you now have your own 2nodes Kubernetes cluster up and running. Let me know if what you think and if you have questions or comments.