3d-photo-inpainting
VQGAN-CLIP
3d-photo-inpainting | VQGAN-CLIP | |
---|---|---|
22 | 67 | |
6,828 | 2,563 | |
0.1% | - | |
0.0 | 0.0 | |
8 months ago | over 1 year ago | |
Python | Python | |
GNU General Public License v3.0 or later | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
3d-photo-inpainting
- I have an AI Generated jpg. I want to add subtle looping animation to it
-
Whats the latest and greatest in 3d img2img/txt2img?
If you are looking to create actual 3d models, the DepthMap extension does have a function to create PLY models with vertex color information, and to render clips with simple camera moves from that extracted 3d scene, including inpainting (as per the 3d-photo-inpainting paper)
-
Quick test of AI and Blender with camera projection.
The depthmap extension for A1111 has implemented the 3d-photo-inpainting code that is doing that kind of thing. That's what I used to use, first on a Colab, and then adapted for windows so I could run it locally. But it's much more convenient to do it directly from the Automatic1111 WebUI.
- Is there an extension that does this?
-
Generate multiple complex subjects on a single image all at once with a depth aware custom extension!
But things are even older than stable diffusion.
-
Coronal mass ejection of the sun. Image from r/space. Crossview ML generated
It's a slightly modified version of https://shihmengli.github.io/3D-Photo-Inpainting/
-
[R] META researchers generate realistic renders from unseen views of any human captured from a single-view RGB-D camera
Thanks! I barely did anything though, just took a deep dream'ed photo made by another artist (Daniel Ambrosi) and passed it through this: https://shihmengli.github.io/3D-Photo-Inpainting/ (github and colab at bottom). Didn't even have to come up with the camera trajectory, was one of the presets in the repo
-
Tumultuous Seas
pretty sure it's this: https://github.com/vt-vl-lab/3d-photo-inpainting
- These are the raw frames I got from Gaugan2, but I'll be posting modified versions in the comment section.
- 3D Photography Using Context-Aware Layered Depth Inpainting
VQGAN-CLIP
-
📚 Tutorials & 🎨 AI Art Generation Tool List Mega Thread
VQGAN-CLIP
-
Which is your favorite text to image model overall?
I've screwed with many text-to-image models over the past couple of years, and I found that while I currently enjoy Stable Diffusion's coherency, I have a soft spot for the ImageNet model used by default for VQGAN+CLIP. It easily approaches the uncanny valley when generating people or animals, but makes for great abstract backgrounds and wallpapers. I already have nostalgia for generating images with it on my CPU overnight.
-
Stable Diffusion Announcement
For someone only tangentially familiar with this space, how is this different than e.g. https://github.com/nerdyrodent/VQGAN-CLIP which you can also run at home? Is it the quality of the generated images?
-
Medieval Noir - VQGAN-CLIP - COCO Checkpoint
Used https://github.com/nerdyrodent/VQGAN-CLIP
- Once have access, do you run it on your computer or over the internet on Open-AI's computers?
- How to get AI imaging effect in Premiere pro
-
A Guide to Asking Robots to Design Stained Glass Windows
I don't have any of the DALL-Es but I do have a couple from github [1], [2] which gave these outputs[3]
[1] https://github.com/nerdyrodent/VQGAN-CLIP
-
How not to waste $1600?
If you want to try your hand at buggering your whole system - try playing with AI image generation as it uses all possible computer assets :D . There is a lot of forms and installations for those but I VQGANs from github the easiest. Problem is that some require familarity with shell, python and in some cases - you need to enable the Linux subsystem in Windows (is it called a subsystem? it is not exactly a VM). This one is the easiest to install out of all I tried. But I liked the results of Pixray most but I wrecked it. I use this one nowadays.
- Ask HN: Is there a publicly available (not private beta) text-to-image API?
-
Got a Machine Learning Algorithm to depict Aphex
For those that are interested, I used VQGAN-CLIP, specifically this GitHub repository
What are some alternatives?
cupscale - Image Upscaling GUI based on ESRGAN
CLIP-Guided-Diffusion - Just playing with getting CLIP Guided Diffusion running locally, rather than having to use colab.
image-super-resolution - 🔎 Super-scale your images and run experiments with Residual Dense and Adversarial Networks.
DALLE-mtf - Open-AI's DALL-E for large scale training in mesh-tensorflow.
Real-ESRGAN - Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.
caire - Content aware image resize library
deep-daze - Simple command line tool for text to image generation using OpenAI's CLIP and Siren (Implicit neural representation network). Technique was originally created by https://twitter.com/advadnoun
BoostingMonocularDepth
waifu2x - Image Super-Resolution for Anime-Style Art
sharp - High performance Node.js image processing, the fastest module to resize JPEG, PNG, WebP, AVIF and TIFF images. Uses the libvips library.
stable-diffusion - A latent text-to-image diffusion model