-
Notifications
You must be signed in to change notification settings - Fork 30
AWS: Low timeout for NVMe devices #2484
Comments
Where in the CoreOS code-base would this value be updated? |
@johanneswuerbach i'm trying to test this to see if the aws docs will fix this. I was able to get the nvme_core params in the boot. I had to edit the grub in i'm still not see the drives get picked up. Although i'm trying on the i3 bare.metal.
|
@johanneswuerbach |
updating the grub config to use the nvme defaults required by aws. Should solve the failure to pass status checks. (eventually) This only works on kernel version above 4.15. (core timeout max is 255 for below 4.15) coreos/bugs#2464 coreos/bugs#2484 coreos/bugs#2371 https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes
updating the grub config to use the nvme defaults required by aws. Should solve the failure to pass status checks. (eventually) This only works on kernel version above 4.15. (core timeout max is 255 for below 4.15) coreos/bugs#2464 coreos/bugs#2484 coreos/bugs#2371 https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes 05c1f12
updating the grub config to use the nvme defaults required by aws. Should solve the failure to pass status checks. (eventually) This only works on kernel version above 4.15. (core timeout max is 255 for below 4.15) coreos/bugs#2464 coreos/bugs#2484 coreos/bugs#2371 https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes 05c1f12
updating the grub config to use the nvme defaults required by aws. Should solve the failure to pass status checks. (eventually) This only works on kernel version above 4.15. (core timeout max is 255 for below 4.15) coreos/bugs#2464 coreos/bugs#2484 coreos/bugs#2371 https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes
updating the grub config to use the nvme defaults required by aws. Should solve the failure to pass status checks. (eventually) This only works on kernel version above 4.15. (core timeout max is 255 for below 4.15) coreos/bugs#2464 coreos/bugs#2484 coreos/bugs#2371 https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes
updating the grub config to use the nvme defaults required by aws. Should solve the failure to pass status checks. (eventually) This only works on kernel version above 4.15. (core timeout max is 255 for below 4.15) coreos/bugs#2464 coreos/bugs#2484 coreos/bugs#2371 https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes
We are also hitting this issue, setting timeouts (255 sec) and retries (10) seems helps. Would love to have this "out-of-the-box" in AWS images. |
Closing as duplicate of #2464. |
Reopening per #2464 (comment). |
The NVMe timeout has been changed for EC2 in the current alpha. It will promote to stable in mid-June. |
I'm running a mix of 2079.6 and 2135.4 and neither has the correct setting... What version is meant to have this fix? |
It's in 2135.0.0 and above, but only for new installs. Machines that are upgraded from older releases retain their previous settings. |
Issue Report
Bug
Container Linux Version
Environment
What hardware/cloud provider/hypervisor is being used to run Container Linux?
AWS us-east-1 c5.2xlarge
Expected Behavior
Having an NVMe I/O Operation Timeout configured according to the recommendations from AWS. https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/nvme-ebs-volumes.html#timeout-nvme-ebs-volumes
Actual Behavior
The NVMe timeout defaults to the kernel default of 30 seconds.
Other Information
Might be related to #2371
The text was updated successfully, but these errors were encountered: