[QUESTION/HELP] Installing NVIDIA GPU operator on a k3d cluster #1458
Replies: 2 comments 1 reply
-
Hey! |
Beta Was this translation helpful? Give feedback.
-
Hello @iwilltry42 , thank you for your reply. I modified the helm chart of gpu operator to disable the symllink creator and was able to avoid the above issue. But now I am running into a new error Error: failed to generate container "7ef3f1c73ce130dc46badbdef38af202bd8e52e495b0e285f823dfd320f4288e" spec: failed to generate spec: path "/run/nvidia/driver" is mounted on "/run" but it is not a shared or slave mount Can you tell me if this is an error with my underlying docker infrastructure? Specifically I am not sure how to mount the /run folder as a shared mount |
Beta Was this translation helpful? Give feedback.
-
Hello,
I am trying to install NVIDIA GPU operator on a K3d cluster. I have a GPU cluster setup according to the docs(https://k3d.io/v5.6.3/usage/advanced/cuda/) and am able to access the GPU using any pods created in the cluster.
But when I try to install NVIDIA GPU operator as described in the install guide (https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/latest/getting-started.html#) the driver validator is stuck in a error loop. Here are the logs from the pod
Can you help me solve this issue?
Beta Was this translation helpful? Give feedback.
All reactions