Markup.
This commit is contained in:
parent
56332586bb
commit
778eceba2b
1 changed files with 13 additions and 10 deletions
23
README.md
23
README.md
|
@ -82,20 +82,23 @@ internal-sto4-test-k8sc-1.rut.sunet.se Ready <none> 16d v1.28.7
|
|||
### Rolling upgrade:
|
||||
|
||||
Drain one controller at the time with:
|
||||
kubectl drain internal-sto4-test-k8sc-0.rut.sunet.se --ignore-daemonset
|
||||
After the first node is drained restart the calio controller with:
|
||||
`kubectl rollout restart deployment calico-kube-controllers -n kube-system`
|
||||
|
||||
kubectl drain internal-sto4-test-k8sc-0.rut.sunet.se --ignore-daemonset
|
||||
After the first node is drained restart the calico controller with:
|
||||
|
||||
kubectl rollout restart deployment calico-kube-controllers -n kube-system
|
||||
After that restart the calico-node running on that host by deleting it. It should be automatically recreated by the controller.
|
||||
`kubectl delete pod calico-node-???? -n kube-system`
|
||||
|
||||
kubectl delete pod calico-node-???? -n kube-system
|
||||
Continue with the workers (Including PG nodes):
|
||||
kubectl drain internal-sto4-test-k8sw-0.rut.sunet.se --force --ignore-daemonsets --delete-emptydir-data --disable-eviction
|
||||
`kubectl delete pod calico-node-???? -n kube-system`
|
||||
|
||||
## Calico problems
|
||||
kubectl drain internal-sto4-test-k8sw-0.rut.sunet.se --force --ignore-daemonsets --delete-emptydir-data --disable-eviction
|
||||
kubectl delete pod calico-node-???? -n kube-system ```
|
||||
|
||||
### Calico problems
|
||||
Calico can get in a bad state. Look for problems like `Candidate IP leak handle och too old resource version` in calico-kube-controllers pod. If theese are found calico can be restarted with:
|
||||
|
||||
kubectl rollout restart deployment calico-kube-controllers -n kube-system
|
||||
kubectl rollout restart daemonset calico-node -n kube-system
|
||||
kubectl rollout restart deployment calico-kube-controllers -n kube-system
|
||||
kubectl rollout restart daemonset calico-node -n kube-system
|
||||
|
||||
This will disrupt the whole cluster for a few seconds.
|
||||
This will disrupt the whole cluster for a few seconds.
|
||||
|
|
Loading…
Add table
Reference in a new issue