-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
upgrade-pod keep restarting even though node is successfully upgraded #72
Comments
Hi @Kun483! Thanks for reporting this. May I ask for a broader scope of logs from the control plane provider? We'll start looking into this and report any further findings here. |
So this is the logs from the
Looks like it fails to fetch things from the given endpoint? |
@HomayoonAlimohammadi I think you should be checking the logs from pod named
upgrade pod is internal to SpectroCloud.
Please let me know if any help is needed to get the correct pod logs. |
@sadysnaat Thanks for correcting me. You're right. I'm going to try to reproduce this issue. |
I tried an in-place upgrade on OB40 with Microk8s v1.27 and 3 control plane nodes.
As a further details:
|
@HomayoonAlimohammadi Our environment is fully airgapped and with a total of 5 bare-metal nodes, no LXD involved. Also with Below are the microk8s pre-loaded images and Pre/PostRun commands that we use under the same environment:
|
@HomayoonAlimohammadi |
@Kun483 Thanks for attaching the logs. Looks like not handling an already created |
@Kun483 Would you please provide me with more info/logs of the |
|
Thanks a lot @Kun483! I guess this means that the upgrade was successful and the |
@HomayoonAlimohammadi Thanks in advance! I want to share another error regarding the HA cluster InPlaceUpgrade in MaaS. Observations: When triggered the InPlaceUpgrade, Logs of
|
What steps did you take and what happened:
I deployed a 3 CP 2 Workers cluster in v1.27. Then I triggered the
InPlaceUpgrade
.upgrade-pod
pod is created in the default namespace of the 1st CP node. Then, this CP node is successfully upgraded to v1.28, and then the corresponding machine got upgraded as well. However, after that,upgrade-pod
got stuck inCrashLoopBackOff
->completed
->CrashLoopBackOff
with more than 20 restarts. When describingupgrade-pod
, it gives me the error below:logs show:
What did you expect to happen:
After the first CP node got upgraded to the desired version,
upgrade-pod
should be deleted from that CP node and re-deployed to the next node.Anything else you would like to add:
I think the code for deleting upgrade-pod never got executed during 1InPlaceUpgrade1. cluster-api-control-plane-provider-microk8s/controllers/reconcile.go at ac3d9e3e8da1b9eb9db0424e355018e4b8faa1b6 · canonical/cluster-api-control-plane-provider-microk8s
Env:
MicroK8s Control Plane and Bootstrap provider: v0.6.10
Upgrade from v1.27 to v.1.28
Infra: MaaS bare Metal for both CP and worker nodes
OS: Ubuntu
UpgradeStrategy: InPlaceUpgrade
The text was updated successfully, but these errors were encountered: