-
Notifications
You must be signed in to change notification settings - Fork 403
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feat/3347 feature add unification support for the rankingquestion #3364
Feat/3347 feature add unification support for the rankingquestion #3364
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Overall LGTM 👍🏻 Just some minor comments and some things to clarify in order to have more context!
Hi! As I have limited bandwidth but this is a key feature for Reward Modeling + RLHF. Could someone add an example (with code snippets) to see what's the expected output of this prepare for training method? Specifically I'm interested and want to double check if we're making it easy or targeting to prepare data for Reward Modeling, i.e., creating pairs of |
Hi Dani, I think this is not directly related to this part. We should add additional class methods to the TrainingTaskMapping. For now we only support .for_text_classification(), but we should also add things like .for_reward_modelling() given certain frameworks. Similarly other tasks ought to be support. If you dear for "ArgillaTrainer" in our issues you should be able yo already find some aligned with the HF mapping. Cheers,DavidEl 10 jul 2023, a las 22:27, Daniel Vila Suero ***@***.***> escribió:
Hi! As I have limited bandwidth but this is a key feature for Reward Modeling + RLHF. Could someone add an example (with code snippets) to see what's the expected output of this prepare for training method?
Specifically I'm interested and want to double check if we're making it easy or targeting to prepare data for Reward Modeling, i.e., creating pairs of chosen rejected responses (as we've shown with this tutorial. If this is not tackled yet, we need to prepare a brief spec for supporting it.
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: ***@***.***>
|
Co-authored-by: Alvaro Bartolome <alvaro@argilla.io>
chore: updated changelog chore: added docstrings
@tomaarsen I also made some changes here that might be relevant, but I saw this was not merged yet. |
…-for-the-rankingquestion
Codecov ReportPatch coverage:
Additional details and impacted files@@ Coverage Diff @@
## develop #3364 +/- ##
===========================================
+ Coverage 90.13% 90.45% +0.31%
===========================================
Files 233 243 +10
Lines 12493 13223 +730
===========================================
+ Hits 11261 11961 +700
- Misses 1232 1262 +30
Flags with carried forward coverage won't be shown. Click here to find out more.
☔ View full report in Codecov by Sentry. |
Description
typing.py
totypes.py
to avoid import errorsRankingQuestionStrategy
RankingQuestionUnification
RankingQuestion
support for the.for_text_classification
method for theTrainingTaskMapping
Closes #3347
Type of change
(Please delete options that are not relevant. Remember to title the PR according to the type of change)
How Has This Been Tested
(Please describe the tests that you ran to verify your changes. And ideally, reference
tests
)tests/client/feedback/test_schemas.py:test_ranking_question_strategy
tests/client/feedback/training/test_schemas.py:test_task_mapping_for_text_classification
Checklist