artroom-stable-diffusion
CLIP-Guided-Diffusion
artroom-stable-diffusion | CLIP-Guided-Diffusion | |
---|---|---|
8 | 4 | |
219 | 377 | |
- | - | |
0.0 | 0.0 | |
about 1 year ago | over 1 year ago | |
Python | Python | |
GNU General Public License v3.0 or later | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
artroom-stable-diffusion
-
Easy-to-use local install of Stable Diffusion released
Github Repo: https://github.com/artmamedov/artroom-stable-diffusion
-
Ran an image of the boys through a few different AI models. Here are a few of the better outcomes.
I use Artroom (Alternative download link), mostly because I'm incompetent and it's the easiest one to set up from all of the things I've found.
- Which is your favorite text to image model overall?
-
Anyone have an idea what the issue is here, attempting to run the optimized scripts but keep hitting the same error, no problem running the normal scripts. Thanks.
wild guess after looking at this.
- I’m buying a 12 GB card for this, how big can I expect to be able to go?
-
image2image throwing errors, unsure how to get it to run
I used a new project called Artroom that did everything for me. I already had the 1.4 model downloaded so I just needed to rename it to model.ckpt and put in the right directory. The creator just added experimental image2image support in the 0.3.0 release, but you can only get that on the authors discord channel at the moment.
-
Cats sitting at a table playing poker
Other pictures are from Pic 1's prompt but with varying parameters. Created using Artroom
-
Stable Diffusion One-Click Install Local GUI
You can get latest from: https://github.com/artmamedov/artroom-stable-diffusion/releases
CLIP-Guided-Diffusion
-
Which is your favorite text to image model overall?
Runner-ups are Craiyon (for being more "creative" than SD), Disco Diffusion, minDALL-E, and CLIP Guided Diffusion.
-
Once have access, do you run it on your computer or over the internet on Open-AI's computers?
-clip guided diffusion https://github.com/nerdyrodent/CLIP-Guided-Diffusion
-
how would i go about running disco diffusion locally?
Nerdy Rodent has a Github repo for this; it should work fine from the Anaconda command line: https://github.com/nerdyrodent/CLIP-Guided-Diffusion
-
PLAYING AGAIN (CLIP GUIDED DIFFUSION) (VQGAN + CLIP) (Beksinski)
As far as I understand, VQGAN is not a guided diffusion model. I've been using a slightly tweaked version of https://github.com/nerdyrodent/CLIP-Guided-Diffusion for diffusion. Once you get it set up the interface is pretty much what you might expect:
What are some alternatives?
disco-diffusion
VQGAN-CLIP - Just playing with getting VQGAN+CLIP running locally, rather than having to use colab.
stable-diffusion-webui - Stable Diffusion web UI
DALLE-mtf - Open-AI's DALL-E for large scale training in mesh-tensorflow.
stable-diffusion-webui-docker - Easy Docker setup for Stable Diffusion with user-friendly UI
mindall-e - PyTorch implementation of a 1.3B text-to-image generation model trained on 14 million image-text pairs
big-sleep - A simple command line tool for text to image generation, using OpenAI's CLIP and a BigGAN. Technique was originally created by https://twitter.com/advadnoun
InvokeAI - InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.
vqgan-clip-app - Local image generation using VQGAN-CLIP or CLIP guided diffusion
stable-diffusion - A latent text-to-image diffusion model
feed_forward_vqgan_clip - Feed forward VQGAN-CLIP model, where the goal is to eliminate the need for optimizing the latent space of VQGAN for each input prompt