stable-diffusion
Our great sponsors
stable-diffusion | stable-diffusion | |
---|---|---|
8 | 20 | |
436 | 338 | |
- | - | |
0.0 | 0.0 | |
12 months ago | over 1 year ago | |
Jupyter Notebook | ||
GNU General Public License v3.0 or later | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-diffusion
-
DALL·E Now Available Without Waitlist
No, sorry, but there's a whole bunch of one-click things now, I think?
I'm running it on Windows 10 using (a modified version of) https://github.com/bfirsh/stable-diffusion.git and Anaconda to create the environment from their `environment.yaml` (all of which was done using the normal `cmd` shell). Then to use it, I activate that env from `cmd` and switch into cygwin `bash` to run the `txt2img.py` script (because it's easier to script, etc.)
-
How do I save the arguments for images I create when using the terminal? (Apple M1 Pro)
I am using the bfirsh version. And yes, I run "pyhthon scripts/txt2imp.py" to generate an image.
-
Current canonical way to install Stable Diffusion on Apple Silicon?
Specifically regarding the first option above, I see that the procedure clones the repository from: https://github.com/bfirsh/stable-diffusion.git
-
One-Click Install Stable Diffusion GUI App for M1 Mac. No Dependencies Needed
Just done a run on my 3080 under Windows using https://github.com/bfirsh/stable-diffusion.git and it's about 8 iterations/sec when nothing else is using CPU or GPU.
-
Using the same seed and same prompt is still resulting in two different images?
I've cloned this repository on my M1 Mac: https://github.com/bfirsh/stable-diffusion/tree/apple-silicon-mps-support
-
Run Stable Diffusion on Your M1 Mac’s GPU
Boom - nice. Here's a fork with that: https://github.com/bfirsh/stable-diffusion/tree/lstein
Requirements are "requirements-mac.txt" which'll need subbing in the guide.
We're testing this out with a few people in Discord before shipping to the blog post.
stable-diffusion
- [Machine Learning] [P] Exécutez une diffusion stable sur le GPU de votre M1 Mac
- High-performance image generation using Stable Diffusion in KerasCV
-
Charl-e: “Stable Diffusion on your Mac in 1 click”
SD on an Intel mac with Vega graphics runs pretty well though — I think it ran at something like ~3-5 iterations/s for me, which is decent. I ran either https://github.com/magnusviri/stable-diffusion or https://github.com/lstein/stable-diffusion which have MPS support
-
Stable Diffusion PR optimizes VRAM, generate 576x1280 images with 6 GB VRAM
https://github.com/magnusviri/stable-diffusion/commit/d0b168...
Copying this change fixed seeds on M1 for me.
-
Intel Mac User, How do I start?
You should be able to run it on a CPU. Maybe try this version. If MPS is supported on your Mac you can check this out.
-
[P] Run Stable Diffusion on your M1 Mac’s GPU
A group of open source hackers forked Stable Diffusion on GitHub and optimized the model to run on Apple's M1 chip, enabling images to be generated in ~ 15 seconds (512x512 pixels, 50 diffusion steps).
-
Run Stable Diffusion on Your M1 Mac’s GPU
Magnusviro [0], the original author of the SD M1 repo credited in this article, has merged his fork into the Lstein Stable Diffusion repo [1], and you can now run Lstein fork with M1 as of a few hours ago.
This adds a ton of functionality - GUI, Upscaling & Facial improvements, weighted subprompts etc.
This has been a big undertaking over the last few days, and I highly recommend checking it out.
[0] https://github.com/magnusviri/stable-diffusion
-
How are Mac people using Windows for A.I. stuff?
You can run it on an M1. Using a macbook M1 pro max with 32Gb I get 512x512 in about 50 seconds. use this branch https://github.com/magnusviri/stable-diffusion/tree/apple-mps-support
-
ResolvePackageNotFound
I had this error too, and I tried a ton of things to get cudatoolkit to install, without any luck. This fork has an environment-mac.yml file that actually got it working on my M1 Max: https://github.com/magnusviri/stable-diffusion/tree/apple-silicon-mps-support
-
If I set a seed value and re-run using the exact same settings, should I get the same image back each time?
But when I run it (locally, using the Mac M1 port), every time I run it creates a different image.
What are some alternatives?
stable_diffusion.openvino
openvino - OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
tvm - Open deep learning compiler stack for cpu, gpu and specialized accelerators
stable-diffusion-webui-docker - Easy Docker setup for Stable Diffusion with user-friendly UI
sd-webui-colab - A repo for the maintenance of the Colab version of stable-diffusion-webui repo
stable-diffusion-webui - Stable Diffusion web UI [Moved to: https://github.com/sd-webui/stable-diffusion-webui]
stable-diffusion - This version of CompVis/stable-diffusion features an interactive command-line script that combines text2img and img2img functionality in a "dream bot" style interface, a WebGUI, and multiple features and other enhancements. [Moved to: https://github.com/invoke-ai/InvokeAI]
Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration
stable-diffusion - A latent text-to-image diffusion model
rocm-build - build scripts for ROCm
invisible-watermark - python library for invisible image watermark (blind image watermark)