Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama : Added support for SmolLm pre-tokenizer (#8608) #8609

Merged
merged 5 commits into from
Jul 22, 2024

Conversation

Stillerman
Copy link
Contributor

@Stillerman Stillerman commented Jul 20, 2024

Adding SmolLM pre-tokenizer support for SmolLM models.

Added tokenizer type for SmolLM-135M in convert-hf-to-gguf-update.py
Added the chkhsh for SmolLM-135M in convert-hf-to-gguf.py
Added LLAMA_VOCAB_PRE_TYPE_SMOLLM enum to llama.h

Ran ./tests/test-tokenizer-0.sh smollm ./models/ggml-vocab-smollm.gguf and Tests passed.

Thank you @m18coppola for #8579

@loubnabnl @anton-l Does src/llama.cpp look right? Any special settings needed in the tokenizer? The .gguf I created for SmolLM-135M seemed to inference well.

@github-actions github-actions bot added the python python script changes label Jul 20, 2024
convert_hf_to_gguf_update.py Outdated Show resolved Hide resolved
src/llama.cpp Show resolved Hide resolved
src/llama.cpp Show resolved Hide resolved
Stillerman and others added 2 commits July 21, 2024 02:48
Co-authored-by: compilade <git@compilade.net>
Co-authored-by: compilade <git@compilade.net>
models/ggml-vocab-smollm.gguf.inp Outdated Show resolved Hide resolved
@ngxson ngxson linked an issue Jul 21, 2024 that may be closed by this pull request
4 tasks
Copy link
Collaborator

@Vaibhavs10 Vaibhavs10 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Tested this! Seems to work well for me too!

@ggerganov ggerganov merged commit d94c6e0 into ggerganov:master Jul 22, 2024
55 checks passed
@eliebak
Copy link

eliebak commented Jul 22, 2024

Seems to work well for me too @Stillerman, thanks a lot for adding support to our model! 🤗

@9cento
Copy link

9cento commented Jul 26, 2024

The 1.7B model prints out nonsense, isn't that supported? Asking for clarification.

arthw pushed a commit to arthw/llama.cpp that referenced this pull request Jul 27, 2024
* Adding SmolLM Pre Tokenizer

* Update convert_hf_to_gguf_update.py

Co-authored-by: compilade <git@compilade.net>

* Update src/llama.cpp

Co-authored-by: compilade <git@compilade.net>

* handle regex

* removed .inp and out .out ggufs

---------

Co-authored-by: compilade <git@compilade.net>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
python python script changes
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Support for SmolLM
6 participants