Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama 3 BPE tokenization needs improvement #4026

Closed
coder543 opened this issue Apr 29, 2024 · 3 comments
Closed

Llama 3 BPE tokenization needs improvement #4026

coder543 opened this issue Apr 29, 2024 · 3 comments
Labels
bug Something isn't working

Comments

@coder543
Copy link

What is the issue?

This PR just merged on llama.cpp, which contained important improvements to how tokenization worked for Llama 3 and other models. An example of the issue is noted here.

Hopefully ollama can update to the latest llama.cpp quickly and make a new release.

OS

Linux

GPU

Nvidia

CPU

AMD

Ollama version

all versions up to this point

@coder543 coder543 added the bug Something isn't working label Apr 29, 2024
@MoonRide303
Copy link

MoonRide303 commented Apr 29, 2024

You might want to wait for ggerganov/llama.cpp#6965 to be merged, too (should happen soon).

@coder543
Copy link
Author

ggerganov/llama.cpp#6965 has been merged now. I'm unclear when things were fixed in ollama, but I just tested with 0.1.35, and I can't reproduce it anymore. Closing.

@dpublic
Copy link

dpublic commented May 13, 2024

The llama.cpp commit link in ollama is dated 4/30 and ggerganov/llama.cpp#6965 was merged to llama.cpp on 5/9.
So, it doesn't look like this merge was included with the last 0.1.37 ollama release.
Does that mean ollama was changed to handle the previous llama.cpp behavior and a future llama.cpp sync in ollama will change behavior?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants