-
-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: fix CUDA images and update go-llama to use full GPU offloading #618
Conversation
GCC 12 is not compatible with nvcc yet. Fixes: #611 Signed-off-by: mudler <mudler@localai.io>
Signed-off-by: mudler <mudler@localai.io>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I have tested this PR on Linux(x86_64) with docker build -t localai .
and docker build --build-arg BUILD_TYPE=cublas localai .
but it is still not working very well.
Don't get that. What's not working well? CI builds fine. |
I test it again, it works for me now. |
After 6-7 requests i stumbled upon Not sure if its a memory issue on my end, since i push it quite a bit right now and it works reliable on the first few ones |
nvcc is not compatible with GCC 12 yet. golang:1.20 images got upgraded to debian bookworm, that uses GCC-12 by default. This PR downgrades images to use bullseye.
Fixes: #611