-
-
Notifications
You must be signed in to change notification settings - Fork 5.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: Support for Higher than 64 LoRa Ranks #3934
Comments
In my personal understanding, you can quikly support
If testing with rank=128 and rank=256 shows no issues, I feel like you can submit a PR to promote support for this feature. |
I do it but get an error while build from source using My change in code:
But get this error: `Building wheels for collected packages: vllm × Building editable for vllm (pyproject.toml) did not run successfully.
note: This error originates from a subprocess, and is likely not a problem with pip. |
@MrzEsma I made the same modifications to bgmv_config.h as you did. Then I built the source using:
I also encountered the build error,as follow error: function "bgmv_kernel<feat_in,feat_out,in_T,out_T,W_T>(out_T *, const in_T *, const W_T *, const int64_t *, int64_t, int64_t, int64_t, int64_t, int64_t, float) [with feat_in=128, feat_out=128, in_T=nv_half, out_T=nv_bfloat16, W_T=nv_half]" explicitly instantiated more than once I deleted two lines from bgmv_config.h then can build successed |
have you solved it? I also meet this problem. |
#5036 can address this problem. |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
This issue has been automatically closed due to inactivity. Please feel free to reopen if you feel it is still relevant. Thank you! |
🚀 The feature, motivation and pitch
Hello,
I was delighted to see the implementation of the multi LoRa feature and would like to express my gratitude and appreciation for your efforts. However, the LoRa we have developed operates with r=128 and r=256, and currently, it does not work for me, resulting in the following error:
ValueError: max_lora_rank (128) must be one of (8, 16, 32, 64).
I am curious to know if there are any plans to support higher ranks? Is this on the priority list? It's quite crucial for us, and we would greatly appreciate any development in this area.
Thank you.
Alternatives
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: