-
-
Notifications
You must be signed in to change notification settings - Fork 123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cudnn_cnn_infer64_8.dll not located. Please advice. #5
Comments
I'm not entirely sure. Maybe you can check out this issue for potential solutions. microsoft/onnxruntime#18973 The problem happens in the inference stage of the speech recognition module after voice activation detection. Which voice recognition are you using? Are you using faster-whisper? If you are using faster-whisper, maybe check their documentation to see if anything is missing. I should probably add this to my documentation... I'm working on dockerizing this program with Nvidia GPU passthrough, which may potentially solve your problem. |
Cool! Great to hear! Thanks for your reply! I have been stepping through and trying to see which part it got stuck in. Yes I'm using the faster-whisper as default. Apparently it got stuck in the function |
Probably it's because I need to install the specific CUDA and CUDNN for my 2060... Can you please share your CUDA and Cudnn versions? |
Well... I'm using an apple silicon Mac so I don't use cuda. I hadn't actually tried running this project on an Nvidia machine yet |
I created the dockerfile and added some docs in the readme for the Nvidia GPU passthrough container. It uses cuda:11.2.2-cudnn8. However, I haven't had the chance to test it. If you feel stuck on fixing Cuda issues, maybe you can take some inspiration from it or just help me test the Nvidia container, which still has a lot of issues, but they are a different set of issues, I guess... |
I will let you know how that goes! Meanwhile, I'm going to test it on my M1 laptop. I will try out docker. Will let you know! |
update! May not be helpful. I didn't have a chance to try the docker image yet. But your repo works for my 4090 setup on a ubuntu20.04 system where cuda and cudnn are correctly set up. So it's a user error on my windows computer! |
Note that there are cudnn 8 and cudnn 9. The command to install onnxruntime for cuda 11 and 12 are different. See the following for detail: onnxruntime-gpu for cuda 11 need cudnn 8, you will need |
For Windows, maybe you should check your system variable, and make sure that your CUDNN files have been placed in the right folders. For example, I'm using Windows 11 with graphics card RTX4060 Laptop, using CUDA 12.6 and it was installed in "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.6", so my system variable "Path" should contains these:
And I need to download "CUDNN for CUDA 12.x" from https://developer.nvidia.com/rdp/cudnn-archive , it should be a .zip file. After unpacking, I need to do these steps:
Then, restart the project, or may need to restart computer for the configuration to take effect. This is my approach to fixing a similar problem when I encounter it, hoping it can be helpful to you. |
2024-07-06 20:14:52.028 | INFO | asr.asr_with_vad:_process_detected_audio:222 - Detected pause after speech. Processing...
2024-07-06 20:14:52.028 | INFO | asr.asr_with_vad:_process_detected_audio:224 - Stopping listening...
Could not locate cudnn_cnn_infer64_8.dll. Please make sure it is in your library path!
Above is the error code.
Platform: Windows 10
Graphic card: NVIDIA-2060
Cudnn installed using
py -m pip install nvidia-cudnn-cu12
The text was updated successfully, but these errors were encountered: