-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] Cannot build --quantization q3f16_1
model
#1005
Comments
I'm not exactly sure what happened but it seems your codebase has been a bit outdated. Would you like to use the latest TVM Unity? https://llm.mlc.ai/docs/install/tvm.html#option-1-prebuilt-package |
@junrushao I have already reinstall the TVM Unity via the command FYI, it works well when I compile the model to q4f16_0 |
cc @vinx13 can you help looking into this one? |
@vinx13 Would like to know any update on this? Thanks a lot 🙇♂️ |
the regression is caused by apache/tvm#15665 I’m still working on a fix, as a workaround you can also revert it for now |
@vinx13 For the workaround, should I build the tvm from the source code with commit before this apache/tvm#15665? The same error occur after I install the tvm from the revert build and run the model compile script. |
I applied the fix on the head of unity branch |
Cannot see the fix commit 🤔 |
@vinx13 could you also cherry pick this fix to the mlc branch? |
I sent a fix in apache/tvm#15881, it will be cherry-picked once merged |
@vinx13 Have installed the tvm based on your fix commit on dc53a6c29. But the same problem still occur when I try to compile the q3f16. Anything have I done wrong? |
I ran |
It works on my side now! Thx you a lot @vinx13 🔥 |
🐛 Bug
Meet an error while building q3f16_1 vicuna-7b-v1.5, the same error also occur when I try to building q3f16_1 for other models
To Reproduce
python3 -m mlc_llm.build --hf-path lmsys/vicuna-7b-v1.5 --target iphone --max-seq-len 4096 --quantization q3f16_1
Error Message
The text was updated successfully, but these errors were encountered: