stable-diffusion-webui
Our great sponsors
stable-diffusion | stable-diffusion-webui | |
---|---|---|
8 | 104 | |
436 | 5,487 | |
- | - | |
0.0 | 10.0 | |
12 months ago | over 1 year ago | |
Python | ||
GNU General Public License v3.0 or later | GNU Affero General Public License v3.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-diffusion
-
DALL·E Now Available Without Waitlist
No, sorry, but there's a whole bunch of one-click things now, I think?
I'm running it on Windows 10 using (a modified version of) https://github.com/bfirsh/stable-diffusion.git and Anaconda to create the environment from their `environment.yaml` (all of which was done using the normal `cmd` shell). Then to use it, I activate that env from `cmd` and switch into cygwin `bash` to run the `txt2img.py` script (because it's easier to script, etc.)
-
How do I save the arguments for images I create when using the terminal? (Apple M1 Pro)
I am using the bfirsh version. And yes, I run "pyhthon scripts/txt2imp.py" to generate an image.
-
Current canonical way to install Stable Diffusion on Apple Silicon?
Specifically regarding the first option above, I see that the procedure clones the repository from: https://github.com/bfirsh/stable-diffusion.git
-
One-Click Install Stable Diffusion GUI App for M1 Mac. No Dependencies Needed
Just done a run on my 3080 under Windows using https://github.com/bfirsh/stable-diffusion.git and it's about 8 iterations/sec when nothing else is using CPU or GPU.
-
Using the same seed and same prompt is still resulting in two different images?
I've cloned this repository on my M1 Mac: https://github.com/bfirsh/stable-diffusion/tree/apple-silicon-mps-support
-
Run Stable Diffusion on Your M1 Mac’s GPU
Boom - nice. Here's a fork with that: https://github.com/bfirsh/stable-diffusion/tree/lstein
Requirements are "requirements-mac.txt" which'll need subbing in the guide.
We're testing this out with a few people in Discord before shipping to the blog post.
stable-diffusion-webui
-
[Stable Diffusion] Je suis confus Aide? - Comment utilisez-vous LDSR avec SD-Webui?
[https://github.com/sd-webui/stable-diffusion-webui/wiki/installation de numéro(https://github.com/sd-webui/stable-diffusion-webui/wiki/installation)
-
[Stable Diffusion] Quelle est la meilleure interface graphique à installer sur Windows?
https://github.com/sd-webui/stable-diffusion-webui (prend beaucoup à installer)
- Daily General Discussion - October 21, 2022
-
Most popular IA to animate?
you can "animate" with stable diffusion usining text to video https://github.com/nateraw/stable-diffusion-videos or https://github.com/sd-webui/stable-diffusion-webui
-
Automatic1111 removed from pinned guide.
I mentioned Automatic1111 on SD-WEBUI and they deleted the comment. I guess this is why. My installation failed on SD-WEBUI and there was no solution for me. I suspect that's why Automatic1111's fork is so popular. He went above and beyond to make sure people with 1660ti's could run SD flawlessly with all the different tools available.
-
.pt to .ckpt
Any way to convert a .pt model to a .ckpt model? Stable-diffusion-webui only seems to support the second type of file but just renaming them does not work:
-
Flooded district by AI
This is Stable-Diffusion. Here is a UI version https://github.com/sd-webui/stable-diffusion-webui
-
AI image generated using the prompt "Streets of Dunwall"
I dunno about the app. I use this https://github.com/sd-webui/stable-diffusion-webui it's very resource hungry though.
-
NMKD Stable Diffusion GUI 1.5.0 is out! Now with exclusion words, CodeFormer face restoration, model merging and pruning tool, even lower VRAM requirements (4 GB), and a ton of quality-of-life improvements. Details in comments.
Haven't tried this GUI yet. Can anyone chime in about how it compares to Automatic1111's and sd-webui/HLKY's? There are so many good repos out there that it's getting hard to keep track of them all
-
Someone just joined 11 GPUs to the Stable Horde. I just tested: 20 gens @ 1024x1024x50 in 2 minutes! All for free!
Maybe those who joined were not aware that they joined the horde :-)
What are some alternatives?
stable_diffusion.openvino
diffusers-uncensored - Uncensored fork of diffusers
tvm - Open deep learning compiler stack for cpu, gpu and specialized accelerators
onnx - Open standard for machine learning interoperability
sd-webui-colab - A repo for the maintenance of the Colab version of stable-diffusion-webui repo
stable-diffusion-webui - Stable Diffusion web UI
stable-diffusion - This version of CompVis/stable-diffusion features an interactive command-line script that combines text2img and img2img functionality in a "dream bot" style interface, a WebGUI, and multiple features and other enhancements. [Moved to: https://github.com/invoke-ai/InvokeAI]
rocm-build - build scripts for ROCm
invisible-watermark - python library for invisible image watermark (blind image watermark)
Dreambooth-Stable-Diffusion - Implementation of Dreambooth (https://arxiv.org/abs/2208.12242) by way of Textual Inversion (https://arxiv.org/abs/2208.01618) for Stable Diffusion (https://arxiv.org/abs/2112.10752). Tweaks focused on training faces, objects, and styles.
stable-diffusion - A latent text-to-image diffusion model
waifu-diffusion - stable diffusion finetuned on weeb stuff