distil-whisper
whisper-turbo
distil-whisper | whisper-turbo | |
---|---|---|
9 | 11 | |
3,225 | 1,594 | |
7.0% | - | |
8.9 | 8.9 | |
16 days ago | 3 months ago | |
Python | TypeScript | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
distil-whisper
- FLaNK Stack 05 Feb 2024
-
Distil-Whisper: a distilled variant of Whisper that is 6x faster
Training code will be released in the Distil-Whisper repository this week, enabling anyone in the community to distill a Whisper model in their choice of language!
- FLaNK Stack Weekly 06 Nov 2023
-
AI — weekly megathread!
Hugging Face released Distil-Whisper, a distilled version of Whisper that is 6 times faster, 49% smaller, and performs within 1% word error rate (WER) on out-of-distribution evaluation sets [Details].
- Distil-Whisper: distilled version of Whisper that is 6 times faster, 49% smaller
- Distil-Whisper is up to 6x faster than Whisper while performing within 1% Word-Error-Rate on out-of-distribution eval sets
-
Distilling Whisper on 20,000 hours of open-sourced audio data
- GitHub page: https://github.com/huggingface/distil-whisper/tree/main
-
Talk-Llama
Is https://github.com/huggingface/distil-whisper on its way to whisper.cpp?
whisper-turbo
- Whisper Turbo: speech recognition in the browser using WebGPU
-
Show HN: Shadeup – A language that makes WebGPU easier
Even just looking at the ability to accelerate llms in the browser on any device without an installation is awesome
For example: fleetwood.dev has a really cool project that does audio transcription in browser on the GPU: https://whisper-turbo.com/#
- Run Whisper on WebGPU with a few lines of JS
- Run LLMs on my own Mac fast and efficient Only 2 MBs
-
Distil-Whisper: distilled version of Whisper that is 6 times faster, 49% smaller
You'd be surprised how capable old GPUs are! I've had great success with people running Whisper-Turbo in the browser on really old hardware: https://whisper-turbo.com/
- Running Whisper on Rust and WebGPU
-
Workers AI: serverless GPU-powered inference on Cloudflare’s global network
Whisper large is only 1.5B params, why not run it client side with something like https://github.com/FL33TW00D/whisper-turbo
(Disclaimer: I am the author)
- Whisper Turbo – Run Whisper Directly in the Browser with Rust and WebGPU
- Whisper Turbo: transcribe 20x faster than realtime using Rust and WebGPU
What are some alternatives?
WhisperInput - Offline voice input panel & keyboard with punctuation for Android.
faster-whisper - Faster Whisper transcription with CTranslate2
pyvideotrans - Translate the video from one language to another and add dubbing. 将视频从一种语言翻译为另一种语言,并添加配音
vid2cleantxt - Python API & command-line tool to easily transcribe speech-based video files into clean text
whisperX - WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)
willow - Open source, local, and self-hosted Amazon Echo/Google Home competitive Voice Assistant alternative
json-masker - High-performance JSON masker library in Java with no runtime dependencies
discourse-ai
project-2501 - Project 2501 is an open-source AI assistant, written in C++.