-
-
Notifications
You must be signed in to change notification settings - Fork 876
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feat: Add example for Mistral #644
Conversation
@@ -0,0 +1,69 @@ | |||
base_model: mistralai/Mistral-7B-v0.1 | |||
base_model_config: mistralai/Mistral-7B-v0.1 | |||
model_type: MistralForCausalLM |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should be able to use AutoModelForCausalLM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, but I think it would be better to specify for later parsing (aka is_mistral_derived..)
let's use |
this PR should fix the issues we were having w transformers main that we had to pin it in the first place #636 |
also, see #646 |
b6aa8ee
to
de10bd4
Compare
* Feat: Add example for Mistral * chore: turn off flash * chore: add is_mistral_derived_model * chore: update following PR
Todo:
In future PR, allow sample_packing.
Upstream PR: huggingface/transformers#26447