Fine-tune Distil-Whisper for personalization #115
Replies: 2 comments 1 reply
-
Hey @Tejaswgupta, this guide here should answer both your questions: https://github.com/huggingface/distil-whisper/tree/main/training#overview-of-training-methods You can convert any personal dataset to Hugging Face datasets using this guide: https://huggingface.co/docs/datasets/audio_dataset Once done, you can run the fine-tuning code by dropping in your custom dataset in HF format. |
Beta Was this translation helpful? Give feedback.
-
Thanks for the quick reply @sanchit-gandhi , my concern with the methods is would they scale well with the number of users.
Thanks a lot! |
Beta Was this translation helpful? Give feedback.
-
What would be the most efficient and likely the easiest method to fine-tune distil-whisper(or whisper) on personal database to adapt to a user's phonetics and vocabulary. Additionally if anyone knows, how much min data should we expect to have to successfully fine-tune the model.
\cc @sanchit-gandhi @Vaibhavs10
Beta Was this translation helpful? Give feedback.
All reactions