Real-ESRGAN
stable-diffusion
Real-ESRGAN | stable-diffusion | |
---|---|---|
131 | 20 | |
26,111 | 338 | |
- | - | |
2.7 | 0.0 | |
22 days ago | over 1 year ago | |
Python | Jupyter Notebook | |
BSD 3-clause "New" or "Revised" License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Real-ESRGAN
-
AI-Powered Nvidia RTX Video HDR Transforms Standard Video into HDR Video
It's not exactly what you're after, as it's anime specific and you need to process the video yourself (eg disassemble to frames, run the upscaler, then assemble back to a movie file), but Real-ESRGAN is really good:
https://github.com/xinntao/Real-ESRGAN/
It's pretty brilliant for cleaning up very old, low resolution anime.
-
Photorealistic Video Generation with Diffusion Models
Just a note you can run upscaling on your home desktop with Real-ESRGAN:
https://github.com/xinntao/Real-ESRGAN
- What software to use for upscaling anime edits
-
What neural net for SISR?
Maybe Real-ESRGAN is a good fit? Even tho it's a couple of years old
- Cant make concurrent calls to Model
-
Outis my beloved
I'm glad you noticed! I upscaled the icon from the wiki using Real-ESRGAN's 4xplus anime model, then photoshopped out the text. Worked far better than waifu2x.
-
ComicMerge (Beta testing version - SafeTensors)
A: Try using High-res Fix and R-ESRGAN 4x+ Anime6B as upscaler
-
Is there any way to upscale local files permanently using Nvidia's RT VSR?
Maybe try this one https://github.com/xinntao/Real-ESRGAN it may work even better.
-
YOASOBI Idol [3840 x 2160]
Screenshotted from the official music video, upscaled to 4k using a state of the art ML model.
-
Compilation of (almost) all end of chapter panels
Do you happen to remember which chapter has that "scene"? You could also try to enhance it yourself, I did it using Real-ESRGAN, which is really easy to use.
stable-diffusion
- [Machine Learning] [P] Exécutez une diffusion stable sur le GPU de votre M1 Mac
- High-performance image generation using Stable Diffusion in KerasCV
-
Charl-e: “Stable Diffusion on your Mac in 1 click”
SD on an Intel mac with Vega graphics runs pretty well though — I think it ran at something like ~3-5 iterations/s for me, which is decent. I ran either https://github.com/magnusviri/stable-diffusion or https://github.com/lstein/stable-diffusion which have MPS support
-
Stable Diffusion PR optimizes VRAM, generate 576x1280 images with 6 GB VRAM
https://github.com/magnusviri/stable-diffusion/commit/d0b168...
Copying this change fixed seeds on M1 for me.
-
Intel Mac User, How do I start?
You should be able to run it on a CPU. Maybe try this version. If MPS is supported on your Mac you can check this out.
-
[P] Run Stable Diffusion on your M1 Mac’s GPU
A group of open source hackers forked Stable Diffusion on GitHub and optimized the model to run on Apple's M1 chip, enabling images to be generated in ~ 15 seconds (512x512 pixels, 50 diffusion steps).
-
Run Stable Diffusion on Your M1 Mac’s GPU
Magnusviro [0], the original author of the SD M1 repo credited in this article, has merged his fork into the Lstein Stable Diffusion repo [1], and you can now run Lstein fork with M1 as of a few hours ago.
This adds a ton of functionality - GUI, Upscaling & Facial improvements, weighted subprompts etc.
This has been a big undertaking over the last few days, and I highly recommend checking it out.
[0] https://github.com/magnusviri/stable-diffusion
-
How are Mac people using Windows for A.I. stuff?
You can run it on an M1. Using a macbook M1 pro max with 32Gb I get 512x512 in about 50 seconds. use this branch https://github.com/magnusviri/stable-diffusion/tree/apple-mps-support
-
ResolvePackageNotFound
I had this error too, and I tried a ton of things to get cudatoolkit to install, without any luck. This fork has an environment-mac.yml file that actually got it working on my M1 Max: https://github.com/magnusviri/stable-diffusion/tree/apple-silicon-mps-support
-
If I set a seed value and re-run using the exact same settings, should I get the same image back each time?
But when I run it (locally, using the Mac M1 port), every time I run it creates a different image.
What are some alternatives?
ESRGAN - ECCV18 Workshops - Enhanced SRGAN. Champion PIRM Challenge on Perceptual Super-Resolution. The training codes are in BasicSR.
openvino - OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
SwinIR - SwinIR: Image Restoration Using Swin Transformer (official repository)
stable-diffusion-webui-docker - Easy Docker setup for Stable Diffusion with user-friendly UI
GFPGAN - GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.
stable-diffusion-webui - Stable Diffusion web UI [Moved to: https://github.com/sd-webui/stable-diffusion-webui]
BSRGAN - Designing a Practical Degradation Model for Deep Blind Image Super-Resolution (ICCV, 2021) (PyTorch) - We released the training code!
Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration
waifu2x - Image Super-Resolution for Anime-Style Art
rocm-build - build scripts for ROCm
Real-ESRGAN-colab - A Real-ESRGAN model trained on a custom dataset
stable-diffusion - This version of CompVis/stable-diffusion features an interactive command-line script that combines text2img and img2img functionality in a "dream bot" style interface, a WebGUI, and multiple features and other enhancements. [Moved to: https://github.com/invoke-ai/InvokeAI]