stable-diffusion-rocm
cog
Our great sponsors
stable-diffusion-rocm | cog | |
---|---|---|
5 | 20 | |
57 | 7,133 | |
- | 8.2% | |
0.0 | 9.4 | |
about 1 year ago | 7 days ago | |
Dockerfile | Python | |
- | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-diffusion-rocm
-
[D] About the current state of ROCm
Re: stable diffusion https://github.com/AshleyYakeley/stable-diffusion-rocm
-
It's time to upscale FSR 2 even further: Meet FSR 2.1
Very easy actually. This is not officially documented, but with a recent enough kernel you don't have to install anything. You can grab the official rocm container and it'll just work. For example for Stable Diffusion see https://github.com/AshleyYakeley/stable-diffusion-rocm/blob/...
-
Running Stable Diffusion on Your GPU with Less Than 10Gb of VRAM
I had good luck with these directions, which let you run inside a docker container:
https://github.com/AshleyYakeley/stable-diffusion-rocm
I had to make the one line change suggested in issue #3 to get it to run under 8GB.
radeontop suggests 4GB might work.
I also had to add this environment variable to make it work on my unsupported radeon 6600xt:
HSA_OVERRIDE_GFX_VERSION=10.3.0
It takes under two minutes per batch of 5 images with the --turbo option.
(Base OS is manjaro; using the distro's version of docker; not the flatpack docker package.)
If you don't have a GPU, paperspace will rent you an appropriate VM.
-
Run Stable Diffusion on Your M1 Mac’s GPU
I have it working on an RX 6800, used the scripts from this repo[0] to build a docker image that has ROCm drivers and PyTorch installed.
I'm running Ubuntu 22.04 LTS as the host OS, didn't have to touch anything beyond the basic Docker install. Next step is build a new Dockerfile that adds in the Stable Diffusion WebUI.[1]
[0] https://github.com/AshleyYakeley/stable-diffusion-rocm
- Dockerfile for easy use on an AMD GPU
cog
-
AI Grant Traction in OSS Startups
View on GitHub
- Insanely Fast Whisper: Transcribe 300 minutes of audio in less than 98 seconds
-
Talk-Llama
I'm in the same situation. I found this cog project to dockerise ML https://github.com/replicate/cog : you write just one python class and a yaml file, and it takes care of the "CUDA hell" and deps. It even creates a flask app in front of your model.
That helps keep your system clean, but someone with big $s please rewrite pytorch to golang or rust or even nodejs / typescript.
-
Llama 2 – Meta AI
https://github.com/replicate/cog
Our thinking was just that a bunch of folks will want to fine-tune right away, then deploy the fine-tunes, so trying to make that easy... Or even just deploy the models-as-is on their own infra without dealing with CUDA insanity!
-
Handling concurrent requests to ML model API
I have used this tool before: https://github.com/replicate/cog/tree/main
-
Opinions on Cog: Containers for machine learning
Then I discovered Cog: Containers for Machine Learning. Looks like a way more flexible solution to plug in the existing infrastructure: you write your custom code and Cog plugs it in a Docker image with FastAPI, no extra ecosystem complexity added.
-
can someone teach me how to install the new stable diffusion repo?
Highly recommend using cog https://github.com/replicate/cog
- Run Stable Diffusion on Your M1 Mac’s GPU
- replicate/cog: Containers for machine learning
-
Why companies move off Heroku (besides the cost)
Dokku Maintainer here.
Dokku also supports Dockerfiles, Docker Images, Tarballs (similar to heroku slugs), and Cloud Native Buildpacks. I'm also actively working on AWS Lambda support (both for simple usage without much config as well as SAM-based usage) and investigating Replicate's Cog[1] and Railways Nixpacks[2] functionalities for building apps.
There are quite a few options in the OSS space (as well as Commercial offerings from new startups and popular incumbents). It's an interesting space to be in, and its always fun to see how new offerings innovate on existing solutions.
[1] https://github.com/replicate/cog
What are some alternatives?
stable-diffusion
nixpacks - App source + Nix packages + Docker = Image
stable_diffusion.openvino
pytorch_wavelets - Pytorch implementation of 2D Discrete Wavelet (DWT) and Dual Tree Complex Wavelet Transforms (DTCWT) and a DTCWT based ScatterNet
tvm - Open deep learning compiler stack for cpu, gpu and specialized accelerators
piku - The tiniest PaaS you've ever seen. Piku allows you to do git push deployments to your own servers.
3d-ken-burns - an implementation of 3D Ken Burns Effect from a Single Image using PyTorch
heroku-review-app-actions - GitHub action to automate managing review apps on your Heroku account
stable-diffusion - Optimized Stable Diffusion modified to run on lower GPU VRAM
stable-diffusion
memray - Memray is a memory profiler for Python