You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Expected behavior was to start LoRA fine-tuning, but it is not coming as expected.
Actual error:
Some weights of Qwen2ForSequenceClassification were not initialized from the model checkpoint at Qwen/Qwen2-0.5B-Instruct and are newly initialized: ['score.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
/home/z004x2xz/miscExcersises/himanshushukla12/trl/examples/scripts/reward_modeling.py:99: UserWarning: You are using a `task_type` that is different than `SEQ_CLS` for PEFT. This will lead to silent bugs Make sure to pass --lora_task_type SEQ_CLS when using this script with PEFT.
warnings.warn(
Traceback (most recent call last):
File "/home/z004x2xz/miscExcersises/himanshushukla12/trl/examples/scripts/reward_modeling.py", line 112, in <module>
trainer = RewardTrainer(
File "/home/z004x2xz/miscExcersises/himanshushukla12/trl/trl/trainer/reward_trainer.py", line 246, in __init__
super().__init__(
TypeError: Trainer.__init__() got an unexpected keyword argument 'processing_class'
Things I tried to fix the issue:
I tried to change processing_class to tokenizer but it didn't worked
The text was updated successfully, but these errors were encountered:
RuntimeError: CUDA error: device-side assert triggered
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
System Info
Here are the versions I'm using
Information
Tasks
examples
folderReproduction
Steps to Reproduce this behavior:
run the command below in
trl
directory:Expected behavior
Expected behavior was to start LoRA fine-tuning, but it is not coming as expected.
Actual error:
Things I tried to fix the issue:
I tried to change
processing_class
totokenizer
but it didn't workedThe text was updated successfully, but these errors were encountered: