is whisper.cpp's inference time slower than original python implementation? #225
Unanswered
utkarshsurwade
asked this question in
Q&A
Replies: 1 comment 3 replies
-
Shouldn't be the case - in my experiments Probably make clean
make bench
./bench -m models/ggml-small.bin -t 8
Also, would be useful to take a look at the CPU model: |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I conducted a short experiment on 1-hour video (https://www.youtube.com/watch?v=j1Hag4Kbpjs) on google colab cpu regarding the inference time of the two implementations.
ggml-base.bin |
20min 5s32min 29sec (colab cpu)ggml-small.bin |
55min 32s1hour 29min 15sec (colab cpu)ggml-large.bin |
4h 58min 6s9hour 33min 3sec (colab cpu)Base whisper python |
11min 27s18min 23 sec (colab cpu)Small whisper python |
24min 24s47min 10sec (colab cpu)Large whisper python |
47min 41sec4hour 45min 10sec (colab cpu)is whisper.cpp's inference time slower than original python implementation for devices other then apple/mac?
Beta Was this translation helpful? Give feedback.
All reactions