v1.7.3 #2645
Replies: 1 comment
-
This is on Mac M1 8GB RAM. large-v3-turbo gives 5x realtime speed but it hallucinates more often than large-v2 does. So it does offer a speed bump for quantized models as it states. 3x compared to 2.3x. whisper 1.7.2 Total duration of audiobook is 00h:32m:24s Whisper large-v2-q8_0 model transcribed at 2.34x realtime speed whisper 1.7.3 whisper.cpp took 00h:10m:53s Total duration of audiobook is 00h:32m:24s Whisper large-v2-q8_0 model transcribed at 2.98x realtime speed 1.7.3 Total duration of audiobook is 00h:32m:24s Whisper large-v3-turbo model transcribed at 5.09x realtime speed if there was a large-v2-turbo that would be ideal. I don't notice any speed improvement from 1.7.2 to 1.7.3 for turbo model but it didn't claim there was one. Both are around 5x realtime speed. |
Beta Was this translation helpful? Give feedback.
-
Overview
What's Changed
q8_0
models todownload-ggml-model.sh
by @mrienstra in Addq8_0
models todownload-ggml-model.sh
#2589download-ggml-model.sh
by @mrienstra in Fix typo indownload-ggml-model.sh
#2623New Contributors
Full Changelog: v1.7.2...v1.7.3
This discussion was created from the release v1.7.3.
Beta Was this translation helpful? Give feedback.
All reactions