stable-fast
ai-notes
stable-fast | ai-notes | |
---|---|---|
11 | 15 | |
1,018 | 4,775 | |
- | - | |
9.3 | 9.8 | |
about 1 month ago | 6 days ago | |
Python | HTML | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-fast
-
Has anyone managed to get TensorRT working in ComfyUI on Windows?
Download (https://github.com/chengzeyi/stable-fast/releases) and install stable-fast binary, compiled according to your system: pip install stable_fast-0.0.13.post3+torch210cu118-cp310-cp310-win_amd64.whl
- Optimum-NVIDIA - 28x faster inference in just 1 line of code !?
- stable-fast for SD inference: Faster than AITemplate, On par with TensorRT
- [N] stable-fast for SD inference: Faster than AITemplate, On par with TensorRT
- Stable-fast for SD inference: Faster than AITemplate, On par with TensorRT
-
SDXL Turbo: A Real-Time Text-to-Image Generation Model
SDXL and ControlNet are already optimized, if thats what you mean: https://github.com/chengzeyi/stable-fast
(Note the links to various SD compilers).
But the whole field is moving so fast that people aren't even adopting the compilers at large.
-
Getting sub 100ms refresh rate on LCMs
> already compiling
Hmm, well if you mean torch.compile, y'all should still check out stable-fast, which is claiming ~16ms/iter on a 4090:
https://github.com/chengzeyi/stable-fast#rtx-4090-512x512-ba...
-
Generate images fast with SD 1.5 while typing on Gradio
Now combine this with an optimized SD implementation, like:
https://github.com/chengzeyi/stable-fast
Or AITemplate, and you are at 15FPS on a larger consumer GPU. 10 with a controlnet you can use for some motion consistency.
-
S-LoRA: Serving Concurrent LoRA Adapters
Since I am sending you down the rabbit hole anyway, you should check out sfast:
https://github.com/chengzeyi/stable-fast
It's, the most promising "fast" and flexible stable diffusion implementation akin to this paper or vLLM that I know of. It doesn't have as many caveats as other implementations, like AITemplate (which is basically Turing+ and linux only) or torch.compile (basically no support for changing inputs/loras).
-
🚀Announcing stable-fast v0.0.5: Speed Optimization for SDXL, Dynamic CUDA Graph
About 2 weeks ago, I released the stable-fast project, which is a lightweight inference performance optimization framework for HuggingFace Diffusers. It provides best performance while keeping the compilation dynamic and flexible, and supports ControlNet and LoRA seamlessly.
ai-notes
-
Minimal implementation of Mamba, the new LLM architecture, in 1 file of PyTorch
the field just moves fast. I have curated a list of non-hypey writers and youtubers who explain these things for a typical SWE audience if you are interested. https://github.com/swyxio/ai-notes/blob/main/Resources/Good%...
- SDXL Turbo: A Real-Time Text-to-Image Generation Model
-
DeepEval – Unit Testing for LLMs
added to my notes! https://github.com/swyxio/ai-notes/
- ChatGPT Code Interpreter Capabilities
-
Google just released a 100% free learning path on Generative AI with 9 Courses
and here are mine, organized by beginner/intermediate/advanced
https://github.com/swyxio/ai-notes/blob/main/README.md#top-a...
and then you can go into the individual modality specific notes for more reading
- Show HN: Self-host Whisper As a Service with GUI and queueing
-
Show HN: YouTube Summaries Using GPT
there's https://learnprompting.org/
i've also been keeping a popular series of notes https://github.com/sw-yx/ai-notes/blob/main/TEXT_PROMPTS.md
-
Show HN: I reverse prompt engineered every Notion AI feature
Direct link to the source prompts are here: https://github.com/sw-yx/ai-notes/blob/main/Resources/Notion...
- GitHub - sw-yx/prompt-eng: notes for prompt engineering
- My hand-curated list of major distros and forks of Stable Diffusion. Please suggest anything I missed!
What are some alternatives?
Fooocus - Focus on prompting and generating
stable-diffusion-ui - Easiest 1-click way to install and use Stable Diffusion on your computer. Provides a browser UI for generating images from text prompts and images. Just enter your text prompt, and see the generated image. [Moved to: https://github.com/easydiffusion/easydiffusion]
gpt-fast - Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
text2image-gui - Somewhat modular text2image GUI, initially just for Stable Diffusion
TensorRT-LLM - TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
diffusionbee-stable-diffusion-ui - Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.
m1_huggingface_diffusers_demo - Demo of how to get HuggingFace Diffusers working on an M1 Mac
perceiver-pytorch - Implementation of Perceiver, General Perception with Iterative Attention, in Pytorch
stable-diffusion - A latent text-to-image diffusion model
rocm-build - build scripts for ROCm
stable-diffusion-webui - Stable Diffusion web UI [Moved to: https://github.com/Sygil-Dev/sygil-webui]
InvokeAI - InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.