Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Not a bug] Support for fine-tuning codestral-mamba? #87

Open
ciscodoojung opened this issue Aug 5, 2024 · 2 comments
Open

[Not a bug] Support for fine-tuning codestral-mamba? #87

ciscodoojung opened this issue Aug 5, 2024 · 2 comments
Labels
bug Something isn't working

Comments

@ciscodoojung
Copy link

Python Version

n/a

Pip Freeze

n/a

Reproduction Steps

n/a

Expected Behavior

n/a

Additional Context

Hi, first of all, thank you all for your help.

I have been trying to use this code (mistral-finetune) to fine-tune the codestral-mamba model (https://huggingface.co/mistralai/Mamba-Codestral-7B-v0.1) but it seems like the code is not expecting mamba architecture. Is there a plan to support to fine-tune codestral-mamba in this repo? or are there any other tools we can use to fine-tune codestral-mamba?

Your response is greatly appreciated! Thank you!

Suggested Solutions

No response

@ciscodoojung ciscodoojung added the bug Something isn't working label Aug 5, 2024
@acodercat
Copy link

+1

@ciscodoojung
Copy link
Author

Just to update the thread (not related to this repo).

Huggingface transformers v4.44.0 was released yesterday to support codestral-mamba training from HF.
Release: https://github.com/huggingface/transformers/releases/tag/v4.44.0
PR: huggingface/transformers#32080

Hope this helps! Thank you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants