-
Notifications
You must be signed in to change notification settings - Fork 966
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Why is GGML so much faster than PyTorch? #382
Comments
|
Yes, it's most likely due to different beam size, greedy vs beam-search decoder, etc. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Test data on my computer, with TR-3970x and RTX3080Ti
Whisper Medium+PyTorch CPU:
Takes 3 hours for audio per hour
Whisper Medium+PyTorch GPU:
10 min per hour
Whisper Large V2+GGML CPU:
30 min per hour
The text was updated successfully, but these errors were encountered: