update
This commit is contained in:
		
							parent
							
								
									78701624c4
								
							
						
					
					
						commit
						9af9a88fc3
					
				|  | @ -40,7 +40,7 @@ user=ubuntu | ||||||
| interface=eth0 | interface=eth0 | ||||||
| 
 | 
 | ||||||
| # Set the virtual IP address (VIP) | # Set the virtual IP address (VIP) | ||||||
| vip=192.168.1.50 | vip=192.168.3.50 | ||||||
| 
 | 
 | ||||||
| # Array of master nodes | # Array of master nodes | ||||||
| masters=($master2 $master3) | masters=($master2 $master3) | ||||||
|  | @ -55,7 +55,7 @@ all=($master1 $master2 $master3 $worker1 $worker2) | ||||||
| allnomaster1=($master2 $master3 $worker1 $worker2) | allnomaster1=($master2 $master3 $worker1 $worker2) | ||||||
| 
 | 
 | ||||||
| #Loadbalancer IP range | #Loadbalancer IP range | ||||||
| lbrange=192.168.1.61-192.168.1.79 | lbrange=192.168.3.60-192.168.3.80 | ||||||
| 
 | 
 | ||||||
| #ssh certificate name variable | #ssh certificate name variable | ||||||
| certName=id_rsa | certName=id_rsa | ||||||
|  | @ -92,17 +92,6 @@ else | ||||||
|     echo -e " \033[32;5mKubectl already installed\033[0m" |     echo -e " \033[32;5mKubectl already installed\033[0m" | ||||||
| fi | fi | ||||||
| 
 | 
 | ||||||
| # Install Docker to generate manifest and daemonset if not already present |  | ||||||
| if ! command -v docker version &> /dev/null |  | ||||||
| then |  | ||||||
|     echo -e " \033[31;5mDocker not found, installing\033[0m" |  | ||||||
|     curl -fsSL https://get.docker.com -o get-docker.sh |  | ||||||
|     sudo sh get-docker.sh |  | ||||||
|     wait $! |  | ||||||
| else |  | ||||||
|     echo -e " \033[32;5mDocker already installed\033[0m" |  | ||||||
| fi |  | ||||||
| 
 |  | ||||||
| # Create SSH Config file to ignore checking (don't use in production!) | # Create SSH Config file to ignore checking (don't use in production!) | ||||||
| echo "StrictHostKeyChecking no" > ~/.ssh/config | echo "StrictHostKeyChecking no" > ~/.ssh/config | ||||||
| 
 | 
 | ||||||
|  | @ -137,32 +126,23 @@ k3sup install \ | ||||||
| echo -e " \033[32;5mFirst Node bootstrapped successfully!\033[0m" | echo -e " \033[32;5mFirst Node bootstrapped successfully!\033[0m" | ||||||
| 
 | 
 | ||||||
| # Step 2: Install Kube-VIP for HA | # Step 2: Install Kube-VIP for HA | ||||||
| kubectl k3s-ha |  | ||||||
| kubectl apply -f https://kube-vip.io/manifests/rbac.yaml | kubectl apply -f https://kube-vip.io/manifests/rbac.yaml | ||||||
| 
 | 
 | ||||||
| # Step 3: Generate Daemonset with Docker | # Step 3: Download kube-vip | ||||||
| sudo docker run --network host --rm ghcr.io/kube-vip/kube-vip:$KVVERSION manifest daemonset \ | curl -sO https://raw.githubusercontent.com/JamesTurland/JimsGarage/main/Kubernetes/K3S-Deploy/kube-vip | ||||||
|     --interface $interface \ | cat kube-vip | sed 's/$interface/'$interface'/g; s/$vip/'$vip'/g' > $HOME/kube-vip.yaml | ||||||
|     --address $vip \ |  | ||||||
|     --inCluster \ |  | ||||||
|     --taint \ |  | ||||||
|     --controlplane \ |  | ||||||
|     --services \ |  | ||||||
|     --arp \ |  | ||||||
|     --leaderElection | tee $HOME/kube-vip.yaml |  | ||||||
| 
 | 
 | ||||||
| # Step 4: Copy kube-vip.yaml to master1 | # Step 4: Copy kube-vip.yaml to master1 | ||||||
| scp -i ~/.ssh/$certName $HOME/kube-vip.yaml $user@$master1:~/kube-vip.yaml | scp -i ~/.ssh/$certName $HOME/kube-vip.yaml $user@$master1:~/kube-vip.yaml | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # Step 5: Connect to Master1 and move kube-vip.yaml | # Step 5: Connect to Master1 and move kube-vip.yaml | ||||||
| 
 |  | ||||||
| ssh $user@$master1 -i ~/.ssh/$certName <<- EOF | ssh $user@$master1 -i ~/.ssh/$certName <<- EOF | ||||||
|   sudo mkdir -p /var/lib/rancher/k3s/server/manifests |   sudo mkdir -p /var/lib/rancher/k3s/server/manifests | ||||||
|   sudo mv kube-vip.yaml /var/lib/rancher/k3s/server/manifests/kube-vip.yaml |   sudo mv kube-vip.yaml /var/lib/rancher/k3s/server/manifests/kube-vip.yaml | ||||||
| EOF | EOF | ||||||
| 
 | 
 | ||||||
| # Step 6: Add new master nodes (servers) | # Step 6: Add new master nodes (servers) & workers | ||||||
| for newnode in "${masters[@]}"; do | for newnode in "${masters[@]}"; do | ||||||
|   k3sup join \ |   k3sup join \ | ||||||
|     --ip $newnode \ |     --ip $newnode \ | ||||||
|  | @ -177,6 +157,7 @@ for newnode in "${masters[@]}"; do | ||||||
|   echo -e " \033[32;5mMaster node joined successfully!\033[0m" |   echo -e " \033[32;5mMaster node joined successfully!\033[0m" | ||||||
| done | done | ||||||
| 
 | 
 | ||||||
|  | # add workers | ||||||
| for newagent in "${workers[@]}"; do | for newagent in "${workers[@]}"; do | ||||||
|   k3sup join \ |   k3sup join \ | ||||||
|     --ip $newagent \ |     --ip $newagent \ | ||||||
|  | @ -184,28 +165,41 @@ for newagent in "${workers[@]}"; do | ||||||
|     --sudo \ |     --sudo \ | ||||||
|     --k3s-version $k3sVersion \ |     --k3s-version $k3sVersion \ | ||||||
|     --server-ip $master1 \ |     --server-ip $master1 \ | ||||||
|     --ssh-key $HOME/.ssh/$certName |     --ssh-key $HOME/.ssh/$certName \ | ||||||
|  |     --k3s-extra-args "--node-label \"longhorn=true\",\"worker=true\"" | ||||||
|   echo -e " \033[32;5mAgent node joined successfully!\033[0m" |   echo -e " \033[32;5mAgent node joined successfully!\033[0m" | ||||||
| done | done | ||||||
| 
 | 
 | ||||||
| # Step 7: Install kube-vip as network LoadBalancer - Install the kube-vip Cloud Provider | # Step 7: Install kube-vip as network LoadBalancer - Install the kube-vip Cloud Provider | ||||||
| kubectl apply -f https://raw.githubusercontent.com/kube-vip/kube-vip-cloud-provider/main/manifest/kube-vip-cloud-controller.yaml | kubectl apply -f https://raw.githubusercontent.com/kube-vip/kube-vip-cloud-provider/main/manifest/kube-vip-cloud-controller.yaml | ||||||
| 
 | 
 | ||||||
| #IP range for loadbalancer services to use | # Step 8: Install Metallb | ||||||
| kubectl create configmap -n kube-system kubevip --from-literal range-global=$lbrange | kubectl apply -f https://raw.githubusercontent.com/metallb/metallb/v0.12.1/manifests/namespace.yaml | ||||||
|  | kubectl apply -f https://raw.githubusercontent.com/metallb/metallb/v0.13.12/config/manifests/metallb-native.yaml | ||||||
|  | # Download ipAddressPool and configure using lbrange above | ||||||
|  | curl -sO https://raw.githubusercontent.com/JamesTurland/JimsGarage/main/Kubernetes/K3S-Deploy/ipAddressPool | ||||||
|  | cat ipAddressPool | sed 's/$lbrange/'$lbrange'/g' > $HOME/ipAddressPool.yaml | ||||||
| 
 | 
 | ||||||
| # Step 8: Test with Nginx | # Step 9: Test with Nginx | ||||||
| kubectl apply -f https://raw.githubusercontent.com/inlets/inlets-operator/master/contrib/nginx-sample-deployment.yaml -n default | kubectl apply -f https://raw.githubusercontent.com/inlets/inlets-operator/master/contrib/nginx-sample-deployment.yaml -n default | ||||||
| kubectl expose deployment nginx-1 --port=80 --type=LoadBalancer -n default | kubectl expose deployment nginx-1 --port=80 --type=LoadBalancer -n default | ||||||
| 
 | 
 | ||||||
| echo -e " \033[32;5mWaiting 20s for K3S to sync and LoadBalancer to come online\033[0m" | echo -e " \033[32;5mWaiting for K3S to sync and LoadBalancer to come online\033[0m" | ||||||
| 
 | 
 | ||||||
| while [[ $(kubectl get pods -l app=nginx -o 'jsonpath={..status.conditions[?(@.type=="Ready")].status}') != "True" ]]; do | while [[ $(kubectl get pods -l app=nginx -o 'jsonpath={..status.conditions[?(@.type=="Ready")].status}') != "True" ]]; do | ||||||
|    sleep 1 |    sleep 1 | ||||||
| done | done | ||||||
| 
 | 
 | ||||||
|  | # Step 10: Deploy IP Pools and l2Advertisement | ||||||
|  | kubectl wait --namespace metallb-system \ | ||||||
|  |                 --for=condition=ready pod \ | ||||||
|  |                 --selector=component=controller \ | ||||||
|  |                 --timeout=120s | ||||||
|  | kubectl apply -f ipAddressPool.yaml | ||||||
|  | kubectl apply -f https://raw.githubusercontent.com/JamesTurland/JimsGarage/main/Kubernetes/K3S-Deploy/l2Advertisement.yaml | ||||||
|  | 
 | ||||||
| kubectl get nodes | kubectl get nodes | ||||||
| kubectl get svc | kubectl get svc | ||||||
| kubectl get pods --all-namespaces -o wide | kubectl get pods --all-namespaces -o wide | ||||||
| 
 | 
 | ||||||
| echo -e " \033[32;5mHappy Kubing! Access Nginx at EXTERNAL-IP above\033[0m" | echo -e " \033[32;5mHappy Kubing! Access Nginx at EXTERNAL-IP above\033[0m" | ||||||
|  | @ -0,0 +1,244 @@ | ||||||
|  | #!/bin/bash | ||||||
|  | 
 | ||||||
|  | echo -e " \033[33;5m    __  _          _        ___                            \033[0m" | ||||||
|  | echo -e " \033[33;5m    \ \(_)_ __ ___( )__    / _ \__ _ _ __ __ _  __ _  ___  \033[0m" | ||||||
|  | echo -e " \033[33;5m     \ \ | '_ \` _ \/ __|  / /_\/ _\` | '__/ _\` |/ _\` |/ _ \ \033[0m" | ||||||
|  | echo -e " \033[33;5m  /\_/ / | | | | | \__ \ / /_\\  (_| | | | (_| | (_| |  __/ \033[0m" | ||||||
|  | echo -e " \033[33;5m  \___/|_|_| |_| |_|___/ \____/\__,_|_|  \__,_|\__, |\___| \033[0m" | ||||||
|  | echo -e " \033[33;5m                                               |___/       \033[0m" | ||||||
|  | 
 | ||||||
|  | echo -e " \033[36;5m                      ___ _  _____ ___                     \033[0m" | ||||||
|  | echo -e " \033[36;5m                     | _ \ |/ / __|_  )                    \033[0m" | ||||||
|  | echo -e " \033[36;5m                     |   / ' <| _| / /                     \033[0m" | ||||||
|  | echo -e " \033[36;5m                     |_|_\_|\_\___/___|                    \033[0m" | ||||||
|  | echo -e " \033[36;5m                                                           \033[0m" | ||||||
|  | echo -e " \033[32;5m             https://youtube.com/@jims-garage              \033[0m" | ||||||
|  | echo -e " \033[32;5m                                                           \033[0m" | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | ############################################# | ||||||
|  | # YOU SHOULD ONLY NEED TO EDIT THIS SECTION # | ||||||
|  | ############################################# | ||||||
|  | 
 | ||||||
|  | # Version of Kube-VIP to deploy | ||||||
|  | KVVERSION="v0.6.3" | ||||||
|  | 
 | ||||||
|  | # Set the IP addresses of the admin, masters, and workers nodes | ||||||
|  | admin=192.168.3.5 | ||||||
|  | master1=192.168.3.21 | ||||||
|  | master2=192.168.3.22 | ||||||
|  | master3=192.168.3.23 | ||||||
|  | worker1=192.168.3.24 | ||||||
|  | worker2=192.168.3.25 | ||||||
|  | 
 | ||||||
|  | # User of remote machines | ||||||
|  | user=ubuntu | ||||||
|  | 
 | ||||||
|  | # Interface used on remotes | ||||||
|  | interface=eth0 | ||||||
|  | 
 | ||||||
|  | # Set the virtual IP address (VIP) | ||||||
|  | vip=192.168.3.50 | ||||||
|  | 
 | ||||||
|  | # Array of all master nodes | ||||||
|  | allmasters=($master1 $master2 $master3) | ||||||
|  | 
 | ||||||
|  | # Array of master nodes | ||||||
|  | masters=($master2 $master3) | ||||||
|  | 
 | ||||||
|  | # Array of worker nodes | ||||||
|  | workers=($worker1 $worker2) | ||||||
|  | 
 | ||||||
|  | # Array of all | ||||||
|  | all=($master1 $master2 $master3 $worker1 $worker2) | ||||||
|  | 
 | ||||||
|  | # Array of all minus master1 | ||||||
|  | allnomaster1=($master2 $master3 $worker1 $worker2) | ||||||
|  | 
 | ||||||
|  | #Loadbalancer IP range | ||||||
|  | lbrange=192.168.3.60-192.168.3.80 | ||||||
|  | 
 | ||||||
|  | #ssh certificate name variable | ||||||
|  | certName=id_rsa | ||||||
|  | 
 | ||||||
|  | ############################################# | ||||||
|  | #            DO NOT EDIT BELOW              # | ||||||
|  | ############################################# | ||||||
|  | # For testing purposes - in case time is wrong due to VM snapshots | ||||||
|  | sudo timedatectl set-ntp off | ||||||
|  | sudo timedatectl set-ntp on | ||||||
|  | 
 | ||||||
|  | # Move SSH certs to ~/.ssh and change permissions | ||||||
|  | cp /home/$user/{$certName,$certName.pub} /home/$user/.ssh | ||||||
|  | chmod 600 /home/$user/.ssh/$certName  | ||||||
|  | chmod 644 /home/$user/.ssh/$certName.pub | ||||||
|  | 
 | ||||||
|  | # Install Kubectl if not already present | ||||||
|  | if ! command -v kubectl version &> /dev/null | ||||||
|  | then | ||||||
|  |     echo -e " \033[31;5mKubectl not found, installing\033[0m" | ||||||
|  |     curl -LO "https://dl.k8s.io/release/$(curl -L -s https://dl.k8s.io/release/stable.txt)/bin/linux/amd64/kubectl" | ||||||
|  |     sudo install -o root -g root -m 0755 kubectl /usr/local/bin/kubectl | ||||||
|  | else | ||||||
|  |     echo -e " \033[32;5mKubectl already installed\033[0m" | ||||||
|  | fi | ||||||
|  | 
 | ||||||
|  | # Create SSH Config file to ignore checking (don't use in production!) | ||||||
|  | echo "StrictHostKeyChecking no" > ~/.ssh/config | ||||||
|  | 
 | ||||||
|  | #add ssh keys for all nodes | ||||||
|  | for node in "${all[@]}"; do | ||||||
|  |   ssh-copy-id $user@$node | ||||||
|  | done | ||||||
|  | 
 | ||||||
|  | # Step 1: Create Kube VIP | ||||||
|  | # create RKE2's self-installing manifest dir | ||||||
|  | sudo mkdir -p /var/lib/rancher/rke2/server/manifests | ||||||
|  | # Install the kube-vip deployment into rke2's self-installing manifest folder | ||||||
|  | curl -sL https://raw.githubusercontent.com/JamesTurland/JimsGarage/main/Kubernetes/RKE2/k3s |  vipAddress=$vip vipInterface=$interface sh | sudo tee /var/lib/rancher/rke2/server/manifests/kube-vip.yaml | ||||||
|  | # Find/Replace all k3s entries to represent rke2 | ||||||
|  | sudo sed -i 's/k3s/rke2/g' /var/lib/rancher/rke2/server/manifests/kube-vip.yaml | ||||||
|  | # copy kube-vip.yaml to home directory | ||||||
|  | sudo cp /var/lib/rancher/rke2/server/manifests/kube-vip.yaml ~/kube-vip.yaml | ||||||
|  | # change owner | ||||||
|  | sudo chown $user:$user kube-vip.yaml | ||||||
|  | # make kube folder to run kubectl later | ||||||
|  | mkdir ~/.kube | ||||||
|  | 
 | ||||||
|  | # create the rke2 config file | ||||||
|  | sudo mkdir -p /etc/rancher/rke2 | ||||||
|  | touch config.yaml | ||||||
|  | echo "tls-san:" >> config.yaml  | ||||||
|  | echo "  - $vip" >> config.yaml | ||||||
|  | echo "  - $master1" >> config.yaml | ||||||
|  | echo "  - $master2" >> config.yaml | ||||||
|  | echo "  - $master3" >> config.yaml | ||||||
|  | echo "write-kubeconfig-mode: 0644" >> config.yaml | ||||||
|  | echo "disable:" >> config.yaml | ||||||
|  | echo "  - rke2-ingress-nginx" >> config.yaml | ||||||
|  | # copy config.yaml to rancher directory | ||||||
|  | sudo cp ~/config.yaml /etc/rancher/rke2/config.yaml | ||||||
|  | 
 | ||||||
|  | # update path with rke2-binaries | ||||||
|  | echo 'export KUBECONFIG=/etc/rancher/rke2/rke2.yaml' >> ~/.bashrc ; echo 'export PATH=${PATH}:/var/lib/rancher/rke2/bin' >> ~/.bashrc ; echo 'alias k=kubectl' >> ~/.bashrc ; source ~/.bashrc ; | ||||||
|  | 
 | ||||||
|  | # Step 2: Copy kube-vip.yaml and certs to all masters | ||||||
|  | for newnode in "${allmasters[@]}"; do | ||||||
|  |   scp -i ~/.ssh/$certName $HOME/kube-vip.yaml $user@$newnode:~/kube-vip.yaml | ||||||
|  |   scp -i ~/.ssh/$certName $HOME/config.yaml $user@$newnode:~/config.yaml | ||||||
|  |   scp -i ~/.ssh/$certName ~/.ssh/{$certName,$certName.pub} $user@$newnode:~/.ssh | ||||||
|  |   echo -e " \033[32;5mCopied successfully!\033[0m" | ||||||
|  | done | ||||||
|  | 
 | ||||||
|  | # Step 3: Connect to Master1 and move kube-vip.yaml and config.yaml. Then install RKE2, copy token back to admin machine. We then use the token to bootstrap additional masternodes | ||||||
|  | ssh -tt $user@$master1 -i ~/.ssh/$certName sudo su <<EOF | ||||||
|  | mkdir -p /var/lib/rancher/rke2/server/manifests | ||||||
|  | mv kube-vip.yaml /var/lib/rancher/rke2/server/manifests/kube-vip.yaml | ||||||
|  | mkdir -p /etc/rancher/rke2 | ||||||
|  | mv config.yaml /etc/rancher/rke2/config.yaml | ||||||
|  | echo 'export KUBECONFIG=/etc/rancher/rke2/rke2.yaml' >> ~/.bashrc ; echo 'export PATH=${PATH}:/var/lib/rancher/rke2/bin' >> ~/.bashrc ; echo 'alias k=kubectl' >> ~/.bashrc ; source ~/.bashrc ; | ||||||
|  | curl -sfL https://get.rke2.io | sh - | ||||||
|  | systemctl enable rke2-server.service | ||||||
|  | systemctl start rke2-server.service | ||||||
|  | echo "StrictHostKeyChecking no" > ~/.ssh/config | ||||||
|  | ssh-copy-id -i /home/$user/.ssh/$certName $user@$admin | ||||||
|  | scp -i /home/$user/.ssh/$certName /var/lib/rancher/rke2/server/token $user@$admin:~/token | ||||||
|  | scp -i /home/$user/.ssh/$certName /etc/rancher/rke2/rke2.yaml $user@$admin:~/.kube/rke2.yaml | ||||||
|  | exit | ||||||
|  | EOF | ||||||
|  | echo -e " \033[32;5mMaster1 Completed\033[0m" | ||||||
|  | 
 | ||||||
|  | # Step 4: Set variable to the token we just extracted, set kube config location | ||||||
|  | token=`cat token` | ||||||
|  | sudo cat ~/.kube/rke2.yaml | sed 's/127.0.0.1/'$master1'/g' > $HOME/.kube/config | ||||||
|  | sudo chown $(id -u):$(id -g) $HOME/.kube/config | ||||||
|  | export KUBECONFIG=${HOME}/.kube/config | ||||||
|  | sudo cp ~/.kube/config /etc/rancher/rke2/rke2.yaml | ||||||
|  | kubectl get nodes | ||||||
|  | 
 | ||||||
|  | # Step 5: Install kube-vip as network LoadBalancer - Install the kube-vip Cloud Provider | ||||||
|  | kubectl apply -f https://kube-vip.io/manifests/rbac.yaml | ||||||
|  | kubectl apply -f https://raw.githubusercontent.com/kube-vip/kube-vip-cloud-provider/main/manifest/kube-vip-cloud-controller.yaml | ||||||
|  | #IP range for loadbalancer services to use | ||||||
|  | kubectl create configmap -n kube-system kubevip --from-literal range-global=$lbrange | ||||||
|  | 
 | ||||||
|  | # Step 6: Add other Masternodes, note we import the token we extracted from step 3 | ||||||
|  | for newnode in "${masters[@]}"; do | ||||||
|  |   ssh -tt $user@$newnode -i ~/.ssh/$certName sudo su <<EOF | ||||||
|  |   mkdir -p /etc/rancher/rke2 | ||||||
|  |   touch /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "token: $token" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "server: https://$master1:9345" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "tls-san:" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "  - $vip" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "  - $master1" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "  - $master2" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "  - $master3" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   curl -sfL https://get.rke2.io | sh - | ||||||
|  |   systemctl enable rke2-server.service | ||||||
|  |   systemctl start rke2-server.service | ||||||
|  |   exit | ||||||
|  | EOF | ||||||
|  |   echo -e " \033[32;5mMaster node joined successfully!\033[0m" | ||||||
|  | done | ||||||
|  | 
 | ||||||
|  | kubectl get nodes | ||||||
|  | 
 | ||||||
|  | # Step 7: Add Workers | ||||||
|  | for newnode in "${workers[@]}"; do | ||||||
|  |   ssh -tt $user@$newnode -i ~/.ssh/$certName sudo su <<EOF | ||||||
|  |   mkdir -p /etc/rancher/rke2 | ||||||
|  |   touch /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "token: $token" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "server: https://$vip:9345" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "node-label:" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "  - worker=true" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   echo "  - longhorn=true" >> /etc/rancher/rke2/config.yaml | ||||||
|  |   curl -sfL https://get.rke2.io | INSTALL_RKE2_TYPE="agent" sh - | ||||||
|  |   systemctl enable rke2-agent.service | ||||||
|  |   systemctl start rke2-agent.service | ||||||
|  |   exit | ||||||
|  | EOF | ||||||
|  |   echo -e " \033[32;5mMaster node joined successfully!\033[0m" | ||||||
|  | done | ||||||
|  | 
 | ||||||
|  | kubectl get nodes | ||||||
|  | 
 | ||||||
|  | # Step 8: Install Rancher (Optional - Delete if not required) | ||||||
|  | #Install Helm | ||||||
|  | curl -fsSL -o get_helm.sh https://raw.githubusercontent.com/helm/helm/main/scripts/get-helm-3 | ||||||
|  | chmod 700 get_helm.sh | ||||||
|  | ./get_helm.sh | ||||||
|  | 
 | ||||||
|  | # Add Rancher Helm Repo & create namespace | ||||||
|  | helm repo add rancher-latest https://releases.rancher.com/server-charts/latest | ||||||
|  | kubectl create namespace cattle-system | ||||||
|  | 
 | ||||||
|  | # Install Cert-Manager | ||||||
|  | kubectl apply -f https://github.com/cert-manager/cert-manager/releases/download/v1.13.2/cert-manager.crds.yaml | ||||||
|  | helm repo add jetstack https://charts.jetstack.io | ||||||
|  | helm repo update | ||||||
|  | helm install cert-manager jetstack/cert-manager \ | ||||||
|  | --namespace cert-manager \ | ||||||
|  | --create-namespace \ | ||||||
|  | --version v1.13.2 | ||||||
|  | kubectl get pods --namespace cert-manager | ||||||
|  | 
 | ||||||
|  | # Install Rancher | ||||||
|  | helm install rancher rancher-latest/rancher \ | ||||||
|  |  --namespace cattle-system \ | ||||||
|  |  --set hostname=rancher.my.org \ | ||||||
|  |  --set bootstrapPassword=admin | ||||||
|  | kubectl -n cattle-system rollout status deploy/rancher | ||||||
|  | kubectl -n cattle-system get deploy rancher | ||||||
|  | 
 | ||||||
|  | # Add Rancher LoadBalancer | ||||||
|  | kubectl get svc -n cattle-system | ||||||
|  | kubectl expose deployment rancher --name=rancher-lb --port=443 --type=LoadBalancer -n cattle-system | ||||||
|  | while [[ $(kubectl get svc -n cattle-system 'jsonpath={..status.conditions[?(@.type=="Pending")].status}') = "True" ]]; do | ||||||
|  |    sleep 5 | ||||||
|  |    echo -e " \033[32;5mWaiting for LoadBalancer to come online\033[0m"  | ||||||
|  | done | ||||||
|  | kubectl get svc -n cattle-system | ||||||
|  | 
 | ||||||
|  | echo -e " \033[32;5mAccess Rancher from the IP above - Password is admin!\033[0m" | ||||||
		Loading…
	
		Reference in New Issue
	
	 James Turland
						James Turland