-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[LoftQConfig + LoraConfig] throws size matmul mismatch error #1240
Comments
Thanks for reporting. Do you get an error like this?
This is because of the bug mentioned here: If I change this line: peft/src/peft/utils/loftq_utils.py Line 201 in fc9f4b3
to: if not is_bnb_4bit_available() or num_bits == 8: I get some progress with your example, but unfortunately encounter another issue:
Similar thing happens when I try to use 4bit instead of 8bit (remember to send the model to cuda). Interestingly, it works for me when using a different architecture (bloomz-560m), both with 4bit and 8bit (when applying the fix above). Therefore, I suspect it's somehow related to the model architecture (we had some issues with gpt2 in the past). Ping @yxli2123 |
Hi, gpt2 uses |
I guess this is the reason. I tested on |
I think the question has been answered, if something new comes up, feel free to re-open. |
I get a similar error on
and my quantization config is
The error is |
It's also happening with |
I see the issue. I was quantization the model on load with |
System Info
pytorch/pytorch:2.1.1-cuda12.1-cudnn8-devel
Who can help?
@pac
Information
Tasks
examples
folderReproduction
I'm testing PEFT Lora Initialization options.
The script I'm testing is official huggingface run_clm.py script and I pasted the lora config part. Nothing else is added or deleted from the original file.
Here is my running script.
Expected behavior
I expected LoftQ Initialization
The text was updated successfully, but these errors were encountered: