openai-whisper-realtime
wer_are_we
openai-whisper-realtime | wer_are_we | |
---|---|---|
1 | 4 | |
180 | 1,862 | |
- | - | |
10.0 | 1.8 | |
over 1 year ago | almost 2 years ago | |
Python | ||
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
openai-whisper-realtime
-
Whisper – open source speech recognition by OpenAI
I tried running it in realtime with live audio input (kind of).
You can find the python script in this repo: https://github.com/tobiashuttinger/openai-whisper-realtime
wer_are_we
-
Lichess Voice Recognition Beta is now Live!
https://github.com/syhw/wer_are_we https://github.com/Franck-Dernoncourt/ASR_benchmark#benchmark-results
-
OpenAI Whisper Model Comparison
Great breakdown… with some interesting results and a ton of effort.
Are there any open benchmarks like this for all models that are actually runnable like the data exposed in https://github.com/syhw/wer_are_we but with some of your additional metrics?
-
Whisper – open source speech recognition by OpenAI
The authors do explicitly state that they're trying to do a lot of fancy new stuff here, like be multilingual, rather than pursuing just accuracy.
[1] https://github.com/syhw/wer_are_we
- This sub is NOT bullying you
What are some alternatives?
whisper - Robust Speech Recognition via Large-Scale Weak Supervision
plaidml - PlaidML is a framework for making deep learning work everywhere.
DeepSpeech-examples - Examples of how to use or integrate DeepSpeech
mycroft-core - Mycroft Core, the Mycroft Artificial Intelligence platform.
NeMo - A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
py-webrtcvad - Python interface to the WebRTC Voice Activity Detector
dragonfly - Speech recognition framework allowing powerful Python-based scripting and extension of Dragon NaturallySpeaking (DNS), Windows Speech Recognition (WSR), Kaldi and CMU Pocket Sphinx
trashbot - Trashbot helper AI assistant
vosk-api - Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node
stable-diffusion - A latent text-to-image diffusion model