You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have been trying to use this code (mistral-finetune) to fine-tune the codestral-mamba model (https://huggingface.co/mistralai/Mamba-Codestral-7B-v0.1) but it seems like the code is not expecting mamba architecture. Is there a plan to support to fine-tune codestral-mamba in this repo? or are there any other tools we can use to fine-tune codestral-mamba?
Your response is greatly appreciated! Thank you!
Suggested Solutions
No response
The text was updated successfully, but these errors were encountered:
Python Version
Pip Freeze
Reproduction Steps
n/a
Expected Behavior
n/a
Additional Context
Hi, first of all, thank you all for your help.
I have been trying to use this code (mistral-finetune) to fine-tune the codestral-mamba model (https://huggingface.co/mistralai/Mamba-Codestral-7B-v0.1) but it seems like the code is not expecting mamba architecture. Is there a plan to support to fine-tune codestral-mamba in this repo? or are there any other tools we can use to fine-tune codestral-mamba?
Your response is greatly appreciated! Thank you!
Suggested Solutions
No response
The text was updated successfully, but these errors were encountered: