Simon Willison on Nostr: Blogged some notes on the new (still MIT licensed) Whisper Turbo model, quietly ...
Blogged some notes on the new (still MIT licensed) Whisper Turbo model, quietly released by OpenAI yesterday
It’s both smaller and 8x faster than their previous Whisper Large https://simonwillison.net/2024/Oct/1/whisper-large-v3-turbo-model/
And you can run it on a Mac with “pip install mlx-whisper” and then:
import mlx_whisper
print(mlx_whisper.transcribe(
"path/to/audio",
path_or_hf_repo="mlx-community/whisper-turbo"
)["text"])
It’s both smaller and 8x faster than their previous Whisper Large https://simonwillison.net/2024/Oct/1/whisper-large-v3-turbo-model/
And you can run it on a Mac with “pip install mlx-whisper” and then:
import mlx_whisper
print(mlx_whisper.transcribe(
"path/to/audio",
path_or_hf_repo="mlx-community/whisper-turbo"
)["text"])