-
Notifications
You must be signed in to change notification settings - Fork 263
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OpinionsQA #1424
OpinionsQA #1424
Conversation
@@ -68,6 +68,9 @@ class AdapterSpec: | |||
# set of training instances. Used to compute error bars. | |||
num_train_trials: int = 1 | |||
|
|||
# Sample train examples or use deterministic |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Make the comment clearer: Sample train examples (as opposed to taking the first few in order)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
But just curious why the sample_train = False mode is needed? In general, it'd be nice not to add more things into AdapterSpec unless they are necessary.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Here, we are providing models bios in context for steerability. I disabled sampling so that I can ensure that we run through all the bios that are in-context examples, rather than just a randomly chosen subset.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sampling is needed so that we can go through all the training examples (which are our steering groups) rather than sampling a random subset of them
src/helm/benchmark/presentation/run_specs_lm_opinions_ai21_steer-bio.conf
Outdated
Show resolved
Hide resolved
src/helm/benchmark/presentation/run_specs_opinions_qa_openai_default.conf
Outdated
Show resolved
Hide resolved
src/helm/benchmark/adaptation/adapters/in_context_learning_adapter.py
Outdated
Show resolved
Hide resolved
src/helm/benchmark/adaptation/adapters/in_context_learning_adapter.py
Outdated
Show resolved
Hide resolved
src/helm/benchmark/presentation/run_specs_opinions_qa_openai_default.conf
Show resolved
Hide resolved
survey_type: str, | ||
num_logprobs: str, | ||
context: str = "None", | ||
num_train_trials: str = "1", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think you can get rid of this, because we have a RunExpander that allows you to set num_train_trials
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the changes - I think this is almost ready - just a few more localized cleanups and documentation, so hopefully it won't be much additional work.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great, thanks! (Left one more comment about the arguments of the run spec.)
Merge the OpinionsQA dataset into HELM.