You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have a problem currently regarding the finetuning of Flaubert on FLUE. I have a model that I re-trained with custom data, so I have new weights for it, and I currently have it as .json file and .bin files. However, when I want to finetune this model on FLUE tasks, they ask me for vocab and codes files from pretraining, that I don't have when using the hugging face library. I see that there is a module to go from XLM to hugging_face, but not the opposite. Is it possible to transform a model under .json and .bin format to get vocab, codes and .pth files ?
Or maybe there is a clever workaround to this problem ?
Many thanks in advance
The text was updated successfully, but these errors were encountered:
I'm really sorry for the very late reply! I had been on leave for several months and have been quite overwhelmed after coming back to work. I'm not sure if my response is helpful to you at the moment but I think HuggingFace datasets repo offers fine-tuning on FLUE, so maybe you don't need to do the conversion?
Please feel free to let me know if there is something else that I can help you with.
Hello,
I have a problem currently regarding the finetuning of Flaubert on FLUE. I have a model that I re-trained with custom data, so I have new weights for it, and I currently have it as .json file and .bin files. However, when I want to finetune this model on FLUE tasks, they ask me for vocab and codes files from pretraining, that I don't have when using the hugging face library. I see that there is a module to go from XLM to hugging_face, but not the opposite. Is it possible to transform a model under .json and .bin format to get vocab, codes and .pth files ?
Or maybe there is a clever workaround to this problem ?
Many thanks in advance
The text was updated successfully, but these errors were encountered: