Skip to content

Large model requires 4-10x time to process long file. Suggestions to improve time? #1747

Discussion options

You must be logged in to vote

The large model (whisper.load_model('large')) performs significantly worse than the documented times to transcription ratios

The documentation doesn't give times to transcription ratios, it gives only relative speed. Thus, however fast the large model happens to run on your particular hardware, the medium model will be ~2x faster than that, the small model ~6x faster, and so on. Of course the large model will run faster on faster hardware, but the medium model will run ~2x faster than that on the same hardware.

How fast does Whisper run on particular hardware? There are too many hardware choices that it would be too expensive to buy them all and test them, but individual users who have …

Replies: 1 comment 5 replies

Comment options

You must be logged in to vote
5 replies
@philippefutureboy
Comment options

@ryanheise
Comment options

@philippefutureboy
Comment options

@ryanheise
Comment options

@philippefutureboy
Comment options

Answer selected by philippefutureboy
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants