Submitted by ggerganov t3_y0nvqu in MachineLearning
upperfloormaster t1_iruyu5a wrote
So you've benchmarked your impl against existing ones, and the results were precisely "very nice performance"% all across the board.
I see.
ggerganov OP t1_irv0gki wrote
Here are some benchmarks that other people did (both vs CPU and vs GPU):
- vs OpenVINO + ONNX on CPU - more than 2x faster
https://github.com/openai/whisper/discussions/208#discussioncomment-3827022
- vs PyTorch (CPU: i7 11800H, GPU: RTX 3080 Laptop):
https://github.com/ggerganov/whisper.cpp/issues/2#issuecomment-1257808576
- whisper.cpp on Xeon processor
https://github.com/ggerganov/whisper.cpp/issues/16
Also, my implementation is focused for performance on M1 chips and it looks like most of the Python frameworks do not support it properly yet, so I cannot make a proper benchmark.
Additionally, my implementation can also run the "large" model on an android phone (Samsung A52) - would be interesting to see how this compares with existing implementations:
https://github.com/ggerganov/whisper.cpp/issues/18#issue-1395784900
Viewing a single comment thread. View all comments