-
Notifications
You must be signed in to change notification settings - Fork 10.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llava 34b model is working with llamacpp? #5266
Comments
Overall if you use the draft PR #5267 you can convert and inference llava-1.6 But at lower quality input images, likely breaking the finetune a bit. Here are 13B and Mistral 7B - as a demo |
ive got some basic quants for 34b generated from a hacked up script. seems to work pretty good. surely the pr mentioned above will improve it further not sure if this kosher to share here or not, but give a go |
This issue should be closed as LLaVA v1.6 support is now merged. |
@mirek190 working! |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Llava 34b model is working with llamacpp?
The text was updated successfully, but these errors were encountered: