InvokeAI
stable_diffusion.openvino | InvokeAI | |
---|---|---|
47 | 239 | |
1,525 | 21,337 | |
- | 1.1% | |
0.8 | 10.0 | |
7 months ago | about 23 hours ago | |
Python | TypeScript | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable_diffusion.openvino
- FLaNK Stack 05 Feb 2024
-
Installing A1111 Stable Diffusion Error
it might be the --xformers flag, try getting rid of that since your not using cuda you wouldn't be able to run it with xformers and you could also try --use-cpu all ... you can also check this out .. https://github.com/bes-dev/stable_diffusion.openvino .. it's probably your best option if your using CPU, which if your PC Graphics are using Intel UHD 620 then you don't have a GPU and an optimized CPU inference would be best to run
- 4 Reasons to Switch to Intel Arc GPUs
-
why is SD not actually using the GPU?
SD can be run on a CPU without a GPU. I know for certain it can be done with OpenVINO. In fact, on some i7s, it will run at around 3 seconds per iteration. There was a reddit SD thread a while back saying it can be done with Automatic111. Also, soe recent threads on problems with AMD GPUs suggest Automatic1111 is using the CPU rather than the intended GPU. (Fortuanely, I have a GPU, so I don't have to deal with it myself!)
-
Slow Performance on RX 6800 XT; Am I Doing Something Wrong or is ROCm Just this Slow?
I'm not actually entirely convinced that it's even using the GPU. Radeontop shows 0% utilization while the images are generating. Additionally, the listed iteration speed should be impossibly slow for any GPU; it says 26.58s/it, which is slower than just running on a CPU.
-
How can i fix it?
iGPU's are in short not supported. There's this repo that may or may not help you, but even if it did I wouldn't expect much.
-
Stable Diffusion Web UI for Intel Arc
You can also run it in windows native with openvino, there is a barebones webui for it as well in one of the forks.Requires setting cpu to gpu in one the files. https://github.com/bes-dev/stable_diffusion.openvino
-
Intel Arc A770 is underperforming in Tom's Hardware Review
In https://github.com/bes-dev/stable_diffusion.openvino/blob/master/stable_diffusion_engine.py
-
So a new benchmark was done for Stable Diffusion on GPU's
" We ended up using three different Stable Diffusion projects for our testing, mostly because no single package worked on every GPU. For Nvidia, we opted for Automatic 1111's webui version(opens in new tab). AMD GPUs were tested using Nod.ai's Shark version(opens in new tab), while for Intel's Arc GPUs we used Stable Diffusion OpenVINO(opens in new tab). "
- Anyone here using Mac?
InvokeAI
-
Stable Diffusion 3
Probably not, since I have no idea what you're talking about. I've just been using the models that InvokeAI (2.3, I only just now saw there's a 3.0) downloads for me [0]. The SD1.5 one is as good as ever, but the SD2 model introduces artifacts on (many, but not all) faces and copyrighted characters.
[0] https://github.com/invoke-ai/InvokeAI
-
AMD Funded a Drop-In CUDA Implementation Built on ROCm: It's Open-Source
I actually used the rocm/pytorch image you also linked.
I'm not sure what you're pointing to with your reference to the Fedora-based images. I'm quite happy with my NixOS install and really don't want to switch to anything else. And as long as I have the correct kernel module, my host OS really shouldn't matter to run any of the images.
And I'm sure it can be made to work with many base images, my point was just that the dependency management around pytorch was in a bad state, where it is extremely easy to break.
> Anyways, hopefully this PR fixes the immediate issue: https://github.com/invoke-ai/InvokeAI/pull/5714/files
It does! At least for me. It is my PR after all ;)
-
Can some expert analyze a github repo and tell us if it's really safe or not?
The data being flagged is not in that github repo, it's fetched from elsewhere and I don't fancy spending time looking for it. The alert is for 'Sirefef!cfg' which has been reported as a false positive with a bunch of other stable diffusion projects (https://www.reddit.com/r/StableDiffusion/comments/101zjec/trojanwin32sirefefcfg_an_apparently_common_false/, https://www.reddit.com/r/StableDiffusion/comments/xmhukb/trojan_in_waifudiffusion_model_file/, https://github.com/invoke-ai/InvokeAI/issues/2773 )
-
What is the most effcient port of SD to mac?
I haven’t tried it recently, but InvokeAI runs on Mac. Invoke. I used to run on my MacBook, but have since gotten a Win laptop.
-
Easy Stable Diffusion XL in your device, offline
There are already a number of local, inference options that are (crucially) open-source, with more robust feature sets.
And if the defense here is "but Auto1111 and Comfy don't have as user-friendly a UI", that's also already covered. https://github.com/invoke-ai/InvokeAI
-
Ask HN: Selfhosted ChatGPT and Stable-diffusion like alternatives?
https://github.com/invoke-ai/InvokeAI should work on your machine. For LLM models, the smaller ones should run using llama.cpp, but I don't think you'll be happy comparing them to ChatGPT.
- 🚀 InvokeAI 3.4 now supports LCM & LCM-LoRAs and much more!
-
Best ai image generator without a nsfw filter?
Stable Diffusion. /r/stablediffusion There are many tutorials on how to set it up locally and use it. InvokeAI is the easiest way to set it up. https://github.com/invoke-ai/InvokeAI
-
What's the best stable diffusion client for base m1 MacBook air?
InvokeAI
- invoke-ai/InvokeAI
What are some alternatives?
stable-diffusion
stable-diffusion-webui - Stable Diffusion web UI
stable-diffusion
stable-diffusion
stable-diffusion-rocm
ControlNet - Let us control diffusion models!
diffusionbee-stable-diffusion-ui - Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.
ComfyUI - The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface.
stable-diffusion - A latent text-to-image diffusion model
dreambooth-gui
stable-diffusion-webui-ipex-arc - A guide to Intel Arc-enabled (maybe) version of @AUTOMATIC1111/stable-diffusion-webui
stable-diffusion - Optimized Stable Diffusion modified to run on lower GPU VRAM