-
feed_forward_vqgan_clip
Feed forward VQGAN-CLIP model, where the goal is to eliminate the need for optimizing the latent space of VQGAN for each input prompt
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Hello, some news. For those who are interested, I released new models (release 0.2) that you could try and you might find them better (depending on the prompt) than the current one(s), also the problem that was mentioned by /u/Wiskkey is less visible (object parts appearing systematically on top-left), but still not 100% solved, there is still a common global structure that can be identified, but it's more centered on the image. The Colab notebook was updated to use the new models.
Related posts
-
[D] Hosting AI Art Generative ML Model
-
A small experiment on how changes in a text prompt may affect output image in a CLIP-based system
-
MultiDiffusion Region Control, a prompt on each mask webui extension is out.
-
Hubble Diffusion with MultiDiffusion
-
MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation