Starting up RHOCP nodes
Prerequisites
Ensure to start up the external storage such as NAS, SAN, or external devices that need to online prior to boot up.
Ensure the external applications such as DNS, Load Balancer, and DHCP are up and running and reachable from cluster.
NOTE:
If any of the components fail to start, see restore section in the link provided in Backup and Restore of RHOCP.
Procedure
- Start the switches.
- Power on all the infrastructure services like DHCP server, DNS, load balancer, and so on.
-
Power on the registry VM (applicable only for disconnected mode of deployment).
- Log in to the registry server.
- Check the status of docker.
systemctl status docker
- Restart the docker service if the status is inactive.
systemctl start docker
- Check the status of the registry container.
docker ps –a
- If the container is in exited state, start the container with below mentioned command:
docker start containerID
- Validate that the docker registry is working.
docker login -u <username> -p <password> https://<registryvm_fqdn_name>:<port>/v2/_catalog
Output:
Login Succeeded
OR
curl -u <username>:<password> -k https://<registryvm_fqdn_name>:<port>/v2/_catalog
-
Power on the Nimble storage nodes.
- Power on the disk shelves and wait till all the disk lights are blinking.
- Power on the array head shelf.
-
Power on all master nodes.
- Log in to a master node using iLO IP address.
- Open the iLO console.
- Click Momentary Press to power on the node and wait until the node comes up.
- Repeat the steps for the second and third master nodes.
-
Start static pods on all master nodes.
for master in ${masters[@]} do echo "==== $master ====" ssh core@$master 'sudo mv -v $(ls /etc/kubernetes/manifests.stop/*) /etc/kubernetes/manifests/ && sudo rmdir /etc/kubernetes/manifests.stop' done
-
Verify all master nodes are in
Ready
state in theoc get nodes
command output. -
Power on all worker nodes.
- Log in to a worker node using ILO IP address.
- Open the iLO console.
- Click Momentary Press to power on the node and wait until the node comes up.
- Repeat the steps for the rest of the worker nodes in the cluster.
-
Check the cluster status.
- Log in to cluster using
kubeadmin
using one of the following methods:export KUBECONFIG=/var/nps/ISO/ign_config/auth/kubeconfig
login -u kubeadmin -p <token-key> --server=https://api.<cluster_domain>:6443
- Access the cluster using the temporary authentication information or
kubeadmin
.export KUBECONFIG=$(pwd)/tmpadmin-kubeconfig oc get nodes
pwd
: Directory where thetmpadmin-kubeconfig
file is created.
- Run the following command to check status of cluster operator:
oc get co oc get nodes
Wait for few seconds till all the worker nodes are in
Ready
state. - Run the following command to verify all the default Pod and daemons are either in
Running
orCompleted
status:oc get pods –-all-namespaces
- Log in to cluster using
-
Remove the temporary authentication information.
NOTE:
Remove temporary admin only in case you are able to access the cluster successfully with
kubeadmin
or withtmp-admin
and the cluster is successfully up and running.oc adm policy remove-cluster-role-from-user cluster-admin -z tmp-admin -n default
oc delete sa tmp-admin -n default