WhisperFusion
returnn-experiments
WhisperFusion | returnn-experiments | |
---|---|---|
3 | 2 | |
1,390 | 152 | |
3.0% | 1.3% | |
8.7 | 6.4 | |
about 2 months ago | 6 months ago | |
Python | Python | |
- | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
WhisperFusion
- FLaNK Stack 05 Feb 2024
- Show HN: WhisperFusion – Ultra-low latency conversations with an AI chatbot
-
WhisperFusion: Ultra-low latency conversations with an AI chatbot
WhisperFusion is fully open-source - https://github.com/collabora/WhisperFusion
returnn-experiments
-
Show HN: WhisperFusion – Ultra-low latency conversations with an AI chatbot
The code is all released already. You find it here: https://github.com/rwth-i6/returnn-experiments/tree/master/2...
This is TensorFlow-based. But I also have another PyTorch-based implementation already, also public (inside our other repo, i6_experiments). It's not so easy currently to set this up, but I'm working on a simpler pipeline in PyTorch.
We don't have the models online yet, but we can upload them later. But I'm not sure how useful they are outside of research, as they are specifically for those research tasks (Librispeech, Tedlium), and probably don't perform too well on other data.
-
Minimal PyTorch re-implementation of GPT
This works for an architecture which has been well tuned and studied before, like LSTM or Transformer.
Once you do research on the model, testing out things, it often tends to become such kwarg monster in many frameworks.
Having everything (relevant) in one file (even in the config file itself with hyper params) allows you to copy the file for every experiment and modify it inplace. This avoids the kwargs mess. But then the config files are very complex, and can become messy in other ways (esp for research projects). Example: https://github.com/rwth-i6/returnn-experiments/blob/master/2...
Such approach makes it much more flexible and does not mess with the baseline code. As you say, it's more like an evolutionary DNA-like approach, where you then tend to do crossovers with other evolved good-performing configs, etc.
What are some alternatives?
WhisperSpeech - An Open Source text-to-speech system built by inverting Whisper.
minGPT - A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
metaflow - :rocket: Build and manage real-life ML, AI, and data science projects with ease!
iris - Transformers are Sample-Efficient World Models. ICLR 2023, notable top 5%.
FLiPStackWeekly - FLaNK AI Weekly covering Apache NiFi, Apache Flink, Apache Kafka, Apache Spark, Apache Iceberg, Apache Ozone, Apache Pulsar, and more...
stable_diffusion.openvino
openvino-ai-plugins-gimp - GIMP AI plugins with OpenVINO Backend
onnxruntime - ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
TornadoVM - TornadoVM: A practical and efficient heterogeneous programming framework for managed languages
openvino_notebooks - 📚 Jupyter notebook tutorials for OpenVINO™
rye - a Hassle-Free Python Experience
FinGPT - FinGPT: Open-Source Financial Large Language Models! Revolutionize 🔥 We release the trained model on HuggingFace.