Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llava 34b model is working with llamacpp? #5266

Closed
mirek190 opened this issue Feb 1, 2024 · 5 comments
Closed

Llava 34b model is working with llamacpp? #5266

mirek190 opened this issue Feb 1, 2024 · 5 comments
Labels
enhancement New feature or request stale

Comments

@mirek190
Copy link

mirek190 commented Feb 1, 2024

Llava 34b model is working with llamacpp?

@mirek190 mirek190 added the enhancement New feature or request label Feb 1, 2024
@cmp-nct
Copy link
Contributor

cmp-nct commented Feb 2, 2024

Overall if you use the draft PR #5267 you can convert and inference llava-1.6

But at lower quality input images, likely breaking the finetune a bit.
I tested the 13B already and it works very well, despite all those shortcommings

Here are 13B and Mistral 7B - as a demo
https://huggingface.co/cmp-nct/llava-1.6-gguf/tree/main

@cjpais
Copy link
Contributor

cjpais commented Feb 2, 2024

ive got some basic quants for 34b generated from a hacked up script. seems to work pretty good.

surely the pr mentioned above will improve it further

hf link

not sure if this kosher to share here or not, but give a go

@svenstaro
Copy link

This issue should be closed as LLaVA v1.6 support is now merged.

@BarfingLemurs
Copy link
Contributor

@mirek190 working!

@github-actions github-actions bot added the stale label Apr 4, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request stale
Projects
None yet
Development

No branches or pull requests

5 participants