Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama3 Tokenizer #4082

Closed
Bearsaerker opened this issue May 1, 2024 · 4 comments
Closed

Llama3 Tokenizer #4082

Bearsaerker opened this issue May 1, 2024 · 4 comments
Labels
bug Something isn't working

Comments

@Bearsaerker
Copy link

What is the issue?

I requanted the llama3 Sauerkraut with the newest release of llama cpp which should have fixed the tokenizer, but when I load the model into Ollama, I still get the wrong output while people using llama cpp get the right one. So I'd say that there is still something buggy in ollama. Here is the Output.
"What is 7777 + 3333?
Let me calculate that for you!

77,777 (first number) + 33,333 (second number) = 111,110

So the answer is 111,110!"

OS

Linux

GPU

Nvidia

CPU

AMD

Ollama version

0.1.32

@Bearsaerker Bearsaerker added the bug Something isn't working label May 1, 2024
@MoonRide303
Copy link

MoonRide303 commented May 1, 2024

Yeah, same here. Freshly converted GGUFs of both original L3 Instruct and various finetunes that give proper answer to this question in llama.cpp (b2776) or koboldcpp (1.64) fail when imported into ollama (0.1.32).

image

@coder543
Copy link

coder543 commented May 1, 2024

Duplicate? #4026

@MoonRide303
Copy link

@coder543 Yeah, seems to be the same thing.

@Bearsaerker
Copy link
Author

Oh yeah duplicate, Im gonna close this then.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants