stable-diffusion
stable-diffusion
Our great sponsors
stable-diffusion | stable-diffusion | |
---|---|---|
142 | 20 | |
2,438 | 338 | |
- | - | |
9.8 | 0.0 | |
over 1 year ago | over 1 year ago | |
Jupyter Notebook | Jupyter Notebook | |
GNU General Public License v3.0 or later | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-diffusion
- [Stable Diffusion] Aide nécessaire à l'augmentation de la taille du fichier maximum sur l'installation locale
- [Machine Learning] [P] Exécutez une diffusion stable sur le GPU de votre M1 Mac
- Its time!
-
Anybody running SD on a Macbook Pro? What are you using and how did you install it?
Yes, you can install it with Python! https://github.com/lstein/stable-diffusion works with macOS, and you can control all the common parameter via their WebUI or CLI :)
-
How do I save the arguments for images I create when using the terminal? (Apple M1 Pro)
I'm using lstein fork ("dream") and when I create an image from the terminal, it also writes back to the terminal like this:
- I Resurrected “Ugly Sonic” with Stable Diffusion Textual Inversion
-
AI Seamless Texture Generator Built-In to Blender
> Whenever I ask for something like ‘seamless tiling xxxxxx’ it kinda sorta gets the idea, but the resulting texture doesn’t quite tile right.
Getting seamless tiling requires more than just have "seamless tiling" in the prompt. It also depends on if the fork you're using has that feature at all.
https://github.com/lstein/stable-diffusion has the feature, but you need to pass it outside the prompt. So if you use the `dream.py` prompt cli, you can pass it `"Hats on the ground" --seamless` and it should be perfectly tilable.
-
Auto SD Workflow - Update 0.2.0 - "Collections", Password Protection, Brand new UI + more
From https://github.com/lstein/stable-diffusion
-
Stable Diffusion GUIs for Apple Silicon
Stable Diffusion Dream Script: This is the original site/script for supporting macOS. I found this soon after Stable Diffusion was publicly released and it was the site which inspired me to try out using Stable Diffusion on a mac. They have a web-based UI (as well as command-line scripts) and a lot of documentation on how to get things working.
-
Still can't believe this technology is real. My talentless 2 minute sketch on the left.
I’m pretty sure it works for M2 as well - basically the newer ARM-based Macs. The instructions to get it working are detailed! https://github.com/lstein/stable-diffusion
stable-diffusion
- [Machine Learning] [P] Exécutez une diffusion stable sur le GPU de votre M1 Mac
- High-performance image generation using Stable Diffusion in KerasCV
-
Charl-e: “Stable Diffusion on your Mac in 1 click”
SD on an Intel mac with Vega graphics runs pretty well though — I think it ran at something like ~3-5 iterations/s for me, which is decent. I ran either https://github.com/magnusviri/stable-diffusion or https://github.com/lstein/stable-diffusion which have MPS support
-
Stable Diffusion PR optimizes VRAM, generate 576x1280 images with 6 GB VRAM
https://github.com/magnusviri/stable-diffusion/commit/d0b168...
Copying this change fixed seeds on M1 for me.
-
Intel Mac User, How do I start?
You should be able to run it on a CPU. Maybe try this version. If MPS is supported on your Mac you can check this out.
-
[P] Run Stable Diffusion on your M1 Mac’s GPU
A group of open source hackers forked Stable Diffusion on GitHub and optimized the model to run on Apple's M1 chip, enabling images to be generated in ~ 15 seconds (512x512 pixels, 50 diffusion steps).
-
Run Stable Diffusion on Your M1 Mac’s GPU
Magnusviro [0], the original author of the SD M1 repo credited in this article, has merged his fork into the Lstein Stable Diffusion repo [1], and you can now run Lstein fork with M1 as of a few hours ago.
This adds a ton of functionality - GUI, Upscaling & Facial improvements, weighted subprompts etc.
This has been a big undertaking over the last few days, and I highly recommend checking it out.
[0] https://github.com/magnusviri/stable-diffusion
-
How are Mac people using Windows for A.I. stuff?
You can run it on an M1. Using a macbook M1 pro max with 32Gb I get 512x512 in about 50 seconds. use this branch https://github.com/magnusviri/stable-diffusion/tree/apple-mps-support
-
ResolvePackageNotFound
I had this error too, and I tried a ton of things to get cudatoolkit to install, without any luck. This fork has an environment-mac.yml file that actually got it working on my M1 Max: https://github.com/magnusviri/stable-diffusion/tree/apple-silicon-mps-support
-
If I set a seed value and re-run using the exact same settings, should I get the same image back each time?
But when I run it (locally, using the Mac M1 port), every time I run it creates a different image.
What are some alternatives?
waifu-diffusion - stable diffusion finetuned on weeb stuff
openvino - OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
taming-transformers - Taming Transformers for High-Resolution Image Synthesis
stable-diffusion-webui-docker - Easy Docker setup for Stable Diffusion with user-friendly UI
stable-diffusion-webui - Stable Diffusion web UI
stable-diffusion-webui - Stable Diffusion web UI [Moved to: https://github.com/sd-webui/stable-diffusion-webui]
diffusers-uncensored - Uncensored fork of diffusers
Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration
txt2imghd - A port of GOBIG for Stable Diffusion
rocm-build - build scripts for ROCm
dream-textures - Stable Diffusion built-in to Blender
sd-akashic - A compendium of informations regarding Stable Diffusion (SD)