distil-whisper
gpt-llm-trainer | distil-whisper | |
---|---|---|
4 | 9 | |
3,825 | 3,225 | |
- | 7.0% | |
5.4 | 8.9 | |
about 2 months ago | 15 days ago | |
Jupyter Notebook | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
gpt-llm-trainer
- FLaNK Stack Weekly 06 Nov 2023
-
Show HN: Fine-tune your own Llama 2 to replace GPT-3.5/4
Very nice, thanks!
Check out what Matt Shumer put together as well: https://github.com/mshumer/gpt-llm-trainer.
I have used his trainer for auto distillation of GPT-4 into GPT3.5 fine tunes, but plan to do the same for Llama as well.
Cheers!
-
[D] Anyone tried gpt-llm-trainer?
Hey guys, so I stumbled upon this Linkedin post, this guy was showing a jupyter notebook on google colab and was explaining step by step how to train your own model to accomplish very specific tasks, and I believe the base model he was using Llama 2 7B Fine tuning version. This is the github link: https://github.com/mshumer/gpt-llm-trainer
- GPT-LLM-Trainer
distil-whisper
- FLaNK Stack 05 Feb 2024
-
Distil-Whisper: a distilled variant of Whisper that is 6x faster
Training code will be released in the Distil-Whisper repository this week, enabling anyone in the community to distill a Whisper model in their choice of language!
- FLaNK Stack Weekly 06 Nov 2023
-
AI — weekly megathread!
Hugging Face released Distil-Whisper, a distilled version of Whisper that is 6 times faster, 49% smaller, and performs within 1% word error rate (WER) on out-of-distribution evaluation sets [Details].
- Distil-Whisper: distilled version of Whisper that is 6 times faster, 49% smaller
- Distil-Whisper is up to 6x faster than Whisper while performing within 1% Word-Error-Rate on out-of-distribution eval sets
-
Distilling Whisper on 20,000 hours of open-sourced audio data
- GitHub page: https://github.com/huggingface/distil-whisper/tree/main
-
Talk-Llama
Is https://github.com/huggingface/distil-whisper on its way to whisper.cpp?
What are some alternatives?
axolotl - Go ahead and axolotl questions
WhisperInput - Offline voice input panel & keyboard with punctuation for Android.
OpenPipe - Turn expensive prompts into cheap fine-tuned models
pyvideotrans - Translate the video from one language to another and add dubbing. 将视频从一种语言翻译为另一种语言,并添加配音
Llama-2-Onnx
vid2cleantxt - Python API & command-line tool to easily transcribe speech-based video files into clean text
trieve - All-in-one infrastructure for building search, recommendations, and RAG. Trieve combines search language models with tools for tuning ranking and relevance.
faster-whisper - Faster Whisper transcription with CTranslate2
open_model_zoo - Pre-trained Deep Learning models and demos (high quality and extremely fast)
json-masker - High-performance JSON masker library in Java with no runtime dependencies
vllm - A high-throughput and memory-efficient inference and serving engine for LLMs
willow - Open source, local, and self-hosted Amazon Echo/Google Home competitive Voice Assistant alternative