whisper-rs
AITemplate
whisper-rs | AITemplate | |
---|---|---|
2 | 37 | |
584 | 4,460 | |
- | 0.7% | |
8.7 | 8.7 | |
7 days ago | about 7 hours ago | |
Rust | Python | |
The Unlicense | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
whisper-rs
-
extract audio/speech as plain text (such as CC) from a video file (.MP4, .MOV)
You might want to look at whisper-rs which are bindings to whisper.cpp: https://github.com/tazz4843/whisper-rs.
- [P] Pure C/C++ port of OpenAI's Whisper
AITemplate
-
Show HN: Shortbread, a web app that helps you create AI comics in minutes
VoltaML is a relatively vanilla diffusers-based backend, so its not a hairy monster to hack like you may have seen with SAI-based UIs.
The AITTemplate code is a lightly modified version of Facebook's example, code, to get rid of small issues like VRAM spikes: https://github.com/facebookincubator/AITemplate/tree/main/ex...
InvokeAI is also diffusers based, but they seem to mess with the pipeline a bit more.
And anyway, all that may be a better reference for interesting features rather than a backend to try and adopt.
-
List of all the ways to improve performance for stable diffusion.
let me know if you discover any more ways to improve SD. I am currently looking into facebooks AITemplate : https://github.com/facebookincubator/AITemplate
- [R] AITemplate Python to AMD compiler {META}
-
Nearly 2x speedup for SD rendering using AITemplate
Link to AITemplate itself: https://github.com/facebookincubator/AITemplate
- Render a neural network into CUDA/HIP code
- Render neural network into CUDA/HIP code
- AITemplate: a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.
- A1111 vs Olive vs AITemplate.
What are some alternatives?
DeepSpeed-MII - MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
stable-diffusion-webui - Stable Diffusion web UI
XNNPACK - High-efficiency floating-point neural network inference operators for mobile, server, and Web
nebuly - The user analytics platform for LLMs
HIP-CPU - An implementation of HIP that works on CPUs, across OSes.
xformers - Hackable and optimized Transformers building blocks, supporting a composable construction.
voltaML - ⚡VoltaML is a lightweight library to convert and run your ML/DL deep learning models in high performance inference runtimes like TensorRT, TorchScript, ONNX and TVM.
stable-diffusion-tensorflow - Stable Diffusion in TensorFlow / Keras
rocm-gfx803
sd_dreambooth_extension