Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adding Qwen2.5 #1834

Open
wants to merge 14 commits into
base: main
Choose a base branch
from
Open

Adding Qwen2.5 #1834

wants to merge 14 commits into from

Conversation

ysjprojects
Copy link

@ysjprojects ysjprojects commented Nov 20, 2024

see #1709

Qwen2.5

0.5B, 1.5B, 3B, 7B, 14B, 32B, 72B

Qwen2.5 Coder

0.5B, 1.5B, 3B, 7B, 14B, 32B

Both base and instruct models.

Motivation:

  • Proven SOTA coding performance, especially on Qwen2.5-Coder series.
  • One of more recent open-source LM releases with decent performance on general benchmarks that is competitive with proprietary models.
  • Notably strong performance on Chinese benchmarks.
  • SOTA model that goes as small as 0.5B, which is very where and will serve many use cases in small LMs.

Copy link
Collaborator

@Andrei-Aksionov Andrei-Aksionov left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hello @ysjprojects 👋

Thanks for the PR.
I believe we wanted to implement Qwen models for quite a while, but never did.

Overall, it looks wonderful.
I just added a couple of nits.

litgpt/prompts.py Outdated Show resolved Hide resolved
tests/test_tokenizer.py Outdated Show resolved Hide resolved
litgpt/model.py Show resolved Hide resolved
@Andrei-Aksionov
Copy link
Collaborator

One more thing: please update the description of the PR with more info about the model.

@Andrei-Aksionov
Copy link
Collaborator

I did a quick check of 0.5B and 1.5B instruct version (with the fix for conversion script).
Don't know about other languages. but in English even 0.5B performs surprisingly well 🙂.

@ysjprojects After you apply the fix that I've mentioned in the comment, I'll be happy to merge the PR.

@ysjprojects
Copy link
Author

One more thing: please update the description of the PR with more info about the model.

is there some specific details that should be included?

@ysjprojects
Copy link
Author

I did a quick check of 0.5B and 1.5B instruct version (with the fix for conversion script). Don't know about other languages. but in English even 0.5B performs surprisingly well 🙂.

@ysjprojects After you apply the fix that I've mentioned in the comment, I'll be happy to merge the PR.

fixed

@Andrei-Aksionov
Copy link
Collaborator

One more thing: please update the description of the PR with more info about the model.

is there some specific details that should be included?

Would be nice if you added why did you decide to add this exact model.
Coding abilities, excellent support of Chinese, ...
Just a bit of context.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants