-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
adds sentiment example for a 20b model #208
Conversation
The documentation is not available anymore as the PR was closed or merged. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for this! Left few nits!
Let's also add a line about it on the docs 💪 Let me know if you want me to do that!
examples/sentiment/scripts/gpt-neox-20b_peft/s03_gpt-neo-20b_sentiment_peft.py
Outdated
Show resolved
Hide resolved
examples/sentiment/scripts/gpt-neox-20b_peft/s03_gpt-neo-20b_sentiment_peft.py
Outdated
Show resolved
Hide resolved
examples/sentiment/scripts/gpt-neox-20b_peft/s03_gpt-neo-20b_sentiment_peft.py
Outdated
Show resolved
Hide resolved
examples/sentiment/scripts/gpt-neox-20b_peft/s03_gpt-neo-20b_sentiment_peft.py
Outdated
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In general happy to merge. Two comments
- left a nit about the naming
- we should add it to the list of examples in the docs
…entiment_peft.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
…entiment_peft.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
…entiment_peft.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
…entiment_peft.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for this!
Hi,there. I am a beginner in the field of NLP and I have been working with the GPT-J model recently. I came across your code for merging adapter layers into the base model's weights in s02_merge_peft_adapter.py , and I have some questions regarding the merging process. From my understanding, after fine-tuning the model with the LORA layer and running this merging code, the LORA layer is replaced with a new randomly initialized linear layer. However, I did not see any indication in the code that the parameters of the LORA layer were inherited by this new linear layer.If this is the case, then it would mean that my previous training of only the LORA layer was pointless. I would be grateful if you could provide me with some clarification on this matter. Thank you very much for your time and help. |
Hi @Kororinpas |
@younesbelkada Thank you for sharing this information with me. I have already checked out the thread you suggested and the original code. Now, my problem is solved. Thanks again! |
This PR adds a sentiment example for a 20b model.
There are 3 scripts: