-
Notifications
You must be signed in to change notification settings - Fork 9.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
convert.py still fails on llama3 8B-Instruct downloaded directly from Meta (Huggingface works) #7339
Comments
I have a similar problem. I merged 2 llama3 8b models with mergekit and i now want to conver them to gguf. This is the output i got: (.venv) PS C:\Users\gsanr\PycharmProjects\llama.cpp> python convert.py penny-dolphin-einstean-llama |
Could it be related to this issue? #7289 |
Have you tried using |
The Meta version doesn't have one, but has a params.json that looks like this and seems to specify similar params. It doesn't list "architectures" though, which is a required key for the convert-hf script:
|
Llama 3 uses the gpt-2 vocab and tiktoken encoder and decoder. The conversion scripts only implemented support for the HF releases. I'm working on streamlining this entire process because converting has become cumbersome and would like a more fluid experience. If I can get the initial stuff ironed out (it's proving challenging), then I'll see if I can get it in there if I have enough time. If not, hopefully have it setup so someone else can easily plug it in and just play it. For now, it's just best to use the hf to gguf script as the official release isn't currently supported due the complicated nature of how BPE is implemented. Also, it looks like it will be moved to examples to reduce confusion since the majority of users are using huggingface. Not sure what the future for convert.py is, but it looks like it will still be kept around which I appreciate. |
I spent 30 hours downloading the Meta versions. |
I have no idea if this will work, but this is what I would try:
|
This issue was closed because it has been inactive for 14 days since being marked as stale. |
I downloaded the llama3 8B Instruct weights directly from the Meta repository (not Huggingface) https://llama.meta.com/llama-downloads. I then tried to run the convert script using the command suggestions that I found in the comments at #6745 and #6819.
tokenizer.model in the contains this. It's definitely not Protobuf, not sure whether it's bpe
I'm running llama.cpp at current master, which is commit 29c60d8. I skimmed the discussion in #6745 and #6920 for a solution, couldn't find one and downloaded the Huggingface version of llama3 8B Instruct instead, which converted without issues. Here are a few of the commands that I tried to run:
The text was updated successfully, but these errors were encountered: