Replies: 1 comment 1 reply
-
hi @rafael844 , this is very interesting. 1 audio file of 58 seconds, medium model, 15 threads (i7-12700H). Both on CPU , precision = int8. WhisperProcess.py (whisperX) took 38 seconds to transcribe. ( whisper-medium model 1.5 GB) Whisper-llamafile.exe took 24 seconds to transcribe. (ggml-medium model 1.5 GB) ====================================================== Whisper-llamafile.exe took 3 seconds to transcribe. (ggml-base model 145 MB) |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
There is a mozzila llamafile implementation to use whisper with CPU that says its very fast. I didnt try it, but you could take a look if its interesting.
https://github.com/Mozilla-Ocho/llamafile/blob/0.8.13/whisper.cpp/doc/index.md
https://www.youtube.com/watch?v=-mRi-B3t6fA
Beta Was this translation helpful? Give feedback.
All reactions