-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Convert_checkpoint.py failed with LLAMA 3.1 8B instruct #2105
Comments
I am also experiencing this issue when running benchmarks. |
sync up, it's fixed on main branch, post v0.11.0 release |
The llama 3.1 is not supported on TRT-LLM 0.11. It is only supported on main branch now. The first commit we support llama 3.1 is |
Discussed with @byshiue , Llama 3.1 models require transformer >= 4.43.0. Maybe a workaround is temporarily |
Is this supported on a released version now? |
Ouch sorry for the noise, I see that on https://github.com/NVIDIA/TensorRT-LLM/releases/tag/v0.12.0 it's supported. |
System Info
Debian 11
Who can help?
@byshiue @juney-nvidia
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Expected behavior
convert_checkpoint.py should work with llama 3.1
actual behavior
additional notes
N/A
The text was updated successfully, but these errors were encountered: