-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
kvm with containerd needs more time to stop #17967
Conversation
/ok-to-test |
kvm2 driver with docker runtime
Times for minikube start: 51.9s 53.7s 51.4s 51.1s 47.2s Times for minikube ingress: 23.6s 25.1s 27.6s 23.6s 26.1s docker driver with docker runtime
Times for minikube start: 25.0s 24.9s 22.3s 24.4s 24.3s Times for minikube ingress: 17.8s 17.8s 18.8s 20.8s 17.8s docker driver with containerd runtime
Times for minikube start: 24.1s 20.5s 20.8s 23.4s 20.4s Times for minikube ingress: 30.3s 31.3s 30.3s 18.4s 31.3s |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code looks good to me, just waiting for the tests
As a side note 1.5 mins seems like a long time for a instance to stop, but it looks like it's just a KVM limitation
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: prezha, spowelljr The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
thanks @spowelljr yes, i saw that occasional behaviour (as shown in the example above) only for kvm+contrainderd combo (probably the contrainerd itself needs that additional time to stop gracefully) - this should eliminate some tests' flakiness in general btw, for this pr: those three netlify tests look stuck and (most of) the other tests were completed hours ago; update: they're unstuck & done now :) |
These are the flake rates of all failed tests.
To see the flake rates of all tests by environment, click here. |
we see tests flake when the kvm driver is used with the containerd container runtime to stop a vm node (example), so we increase the timeout to allow a vm to shut down gracefully
also, the libvirt's
VIR_DOMAIN_SHUTDOWN
status (where the domain is being shut down) has a more suitable mapping to libmachine'sstate.Stopping
instead ofstate.Running
that can be misleading - eg:minikube/cmd/minikube/cmd/status.go
Lines 341 to 342 in 838ea0c
before
after
note:
Successfully stopped node
following theFailed to stop node
is addressed in pr #17965