You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi thanks for your project !
I've been trying to use your work to punctuate some audios in portuguese, but I got stuck with some problems with the Tokenizer
First I got in punctuate.py:
line 84, in init self.tokenizer = self.whisper_tokenizer.tokenizer AttributeError: 'Tokenizer' object has no attribute 'tokenizer'
By removing the .tokenizer, I got another error in punctuate.py:
line 221 tokenizer has no convert ids tokenizer.convert_ids_to_tokens
Do you have any ideia why this is happening?
The text was updated successfully, but these errors were encountered:
The issue you are experiencing is due to a recent change in whisper (openai/whisper#1044), which has replaced Hugging Face's tokenizer with TikToken. I will modify this repository to ensure compatibility with the latest version of Whisper.
In the meantime, as a workaround, you can use the older version of Whisper by running the following command:
pip install openai-whisper==20230308
Thank you for bringing this to my attention and please let me know if you have any further questions or concerns.
Hi thanks for your project !
I've been trying to use your work to punctuate some audios in portuguese, but I got stuck with some problems with the Tokenizer
First I got in punctuate.py:
line 84, in init self.tokenizer = self.whisper_tokenizer.tokenizer AttributeError: 'Tokenizer' object has no attribute 'tokenizer'
By removing the .tokenizer, I got another error in punctuate.py:
line 221 tokenizer has no convert ids tokenizer.convert_ids_to_tokens
Do you have any ideia why this is happening?
The text was updated successfully, but these errors were encountered: