is whisper.cpp's inference time slower than original python implementation? #225
Unanswered
utkarshsurwade
asked this question in
Q&A
Replies: 1 comment 3 replies
-
Shouldn't be the case - in my experiments Probably make clean
make bench
./bench -m models/ggml-small.bin -t 8
Also, would be useful to take a look at the CPU model: |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I conducted a short experiment on 1-hour video (https://www.youtube.com/watch?v=j1Hag4Kbpjs) on google colab cpu regarding the inference time of the two implementations.
ggml-base.bin |
20min 5s32min 29sec (colab cpu)ggml-small.bin |
55min 32s1hour 29min 15sec (colab cpu)ggml-large.bin |
4h 58min 6s9hour 33min 3sec (colab cpu)Base whisper python |
11min 27s18min 23 sec (colab cpu)Small whisper python |
24min 24s47min 10sec (colab cpu)Large whisper python |
47min 41sec4hour 45min 10sec (colab cpu)is whisper.cpp's inference time slower than original python implementation for devices other then apple/mac?
Beta Was this translation helpful? Give feedback.
All reactions