stable-diffusion
stable-diffusion-webui-feature-showcase
stable-diffusion | stable-diffusion-webui-feature-showcase | |
---|---|---|
142 | 33 | |
2,438 | 968 | |
- | - | |
9.8 | 0.0 | |
over 1 year ago | 7 months ago | |
Jupyter Notebook | ||
GNU General Public License v3.0 or later | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-diffusion
- [Stable Diffusion] Aide nécessaire à l'augmentation de la taille du fichier maximum sur l'installation locale
- [Machine Learning] [P] Exécutez une diffusion stable sur le GPU de votre M1 Mac
- Its time!
-
Anybody running SD on a Macbook Pro? What are you using and how did you install it?
Yes, you can install it with Python! https://github.com/lstein/stable-diffusion works with macOS, and you can control all the common parameter via their WebUI or CLI :)
-
How do I save the arguments for images I create when using the terminal? (Apple M1 Pro)
I'm using lstein fork ("dream") and when I create an image from the terminal, it also writes back to the terminal like this:
- I Resurrected “Ugly Sonic” with Stable Diffusion Textual Inversion
-
AI Seamless Texture Generator Built-In to Blender
> Whenever I ask for something like ‘seamless tiling xxxxxx’ it kinda sorta gets the idea, but the resulting texture doesn’t quite tile right.
Getting seamless tiling requires more than just have "seamless tiling" in the prompt. It also depends on if the fork you're using has that feature at all.
https://github.com/lstein/stable-diffusion has the feature, but you need to pass it outside the prompt. So if you use the `dream.py` prompt cli, you can pass it `"Hats on the ground" --seamless` and it should be perfectly tilable.
-
Auto SD Workflow - Update 0.2.0 - "Collections", Password Protection, Brand new UI + more
From https://github.com/lstein/stable-diffusion
-
Stable Diffusion GUIs for Apple Silicon
Stable Diffusion Dream Script: This is the original site/script for supporting macOS. I found this soon after Stable Diffusion was publicly released and it was the site which inspired me to try out using Stable Diffusion on a mac. They have a web-based UI (as well as command-line scripts) and a lot of documentation on how to get things working.
-
Still can't believe this technology is real. My talentless 2 minute sketch on the left.
I’m pretty sure it works for M2 as well - basically the newer ARM-based Macs. The instructions to get it working are detailed! https://github.com/lstein/stable-diffusion
stable-diffusion-webui-feature-showcase
- How to turn anime image to realistic image in stable diffusion?
- [Stable Diffusion] inversion textuelle avec AUTOMATIC1111 webui
- [Ainudes] Comment créer des nus IA ?
- Is there any documentation for Automatic1111 WebUI?
-
Is there a properly comprehensive guide on prompt syntax?
A1111 https://github.com/AUTOMATIC1111/stable-diffusion-webui-feature-showcase
-
Which one is the "official" version
Here's a quick rundown on a few of the most popular ones with links. I started out using CMDR2, which is very easy to get running as a newbie. Then I kind of graduated to NMKD because I wanted something a little more mainstream but still easy to use. Then, I finally decided I was hungry for all the strange and exotic bells and whistles that SD had to offer me, and so I installed Automatic1111. I also wanted something that would work well with my 4GB GTX 1650 laptop card, because that's considered "low ram" and kind of on the edge for running SD- Automatic1111 fit the bill there, too.
-
At your service...
All generations were on the "Berry's Mix" model, which is made by combining NAI-final, Zenith's F111, r34 and SD1.4 according to this recipe. I used 30ish steps when generating images and inpainting, but 70-80 steps when outpainting because I read here that outpainting really benefits from extra steps. When outpainting I would generate 2-4 versions and pick the least broken one, then tidy up with inpainting.
-
What's the name of this feature?
sounds like "outpainting", one of the very first features listed on 1111 repo with some instructions: https://github.com/AUTOMATIC1111/stable-diffusion-webui-feature-showcase
- How do you expand an image? (image to image)
-
Running neural networks locally.
I have no idea what you're talking about. Just get Automatic1111
What are some alternatives?
waifu-diffusion - stable diffusion finetuned on weeb stuff
CogVideo - Text-to-video generation. The repo for ICLR2023 paper "CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers"
taming-transformers - Taming Transformers for High-Resolution Image Synthesis
glid-3-xl-stable - stable diffusion training
stable-diffusion-webui - Stable Diffusion web UI
diffusers-uncensored - Uncensored fork of diffusers
stable-diffusion - Optimized Stable Diffusion modified to run on lower GPU VRAM
txt2imghd - A port of GOBIG for Stable Diffusion
diffusionbee-stable-diffusion-ui - Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.
dream-textures - Stable Diffusion built-in to Blender
StableDiffusion-Windows-GUI