temporal-shift-module
instruct-pix2pix
Our great sponsors
temporal-shift-module | instruct-pix2pix | |
---|---|---|
3 | 21 | |
2,019 | 5,649 | |
0.9% | - | |
3.0 | 0.0 | |
7 months ago | about 1 year ago | |
Python | Python | |
MIT License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
temporal-shift-module
- Stable Video Diffusion
-
Can two-stream networks trained for video action recognition be used for real-time usecases?
My question mostly has to do with optical flow. One of the two-stream networks I'm interested in trying out is TSN-TSM, as there are pre-trained weights available for it on the Assembly101 dataset released a few months ago.
-
I am having a hard time understanding this paper(Temporal shift module). Can some who have read it before or willing to read it explain me better in a more elaborate way?
This is the paper. (https://arxiv.org/abs/1811.08383). Here they are talking about how they can achieve temporal modelling by moving channels, which I assume are the RGB channels across frames. But I am super confused by the lingo. Here is the repo (https://github.com/mit-han-lab/temporal-shift-module). I can't give better rewards except virtual hugs. Thank you.
instruct-pix2pix
-
Stable Video Diffusion
My guess is you're thinking of InstructPix2Pix[1], with prompts like "make the sky green" or "replace the fruits with cake"
[1] https://github.com/timothybrooks/instruct-pix2pix
-
AI image editors with “text to filter” function?
This comes from https://github.com/timothybrooks/instruct-pix2pix, there is also an extension to use it in Automatic1111 Stable diffusion webui.
- [D] NeRF, LeRF, Prolific Dreamer, Neuralangelo, and a lot of other cool NeRF research
-
Was it SD that had the ability to edit a photo using prompts?
InstructPix2Pix
-
Alternate download location for instruct-pix2pix-00-22000.ckpt?
Is there another place I can download the model? I tried downloading the file using the instructions on this page:
-
Using our photoshop plugin for some cool image editing! :D
It comes from https://github.com/timothybrooks/instruct-pix2pix, you can try it out https://huggingface.co/spaces/timbrooks/instruct-pix2pix
-
instruct pix2pix faces always come out messed up. The rest is really good. Any idea how to fix this?
interesting, I've been running it using this: https://github.com/timothybrooks/instruct-pix2pix/blob/main/LICENSE
-
Everybody is always talking about AGI. I'm more curious about using the tools that we have now.
This is already done and it's already been implemented in the most popular web-ui for stable diffusion too. Granted the results aren't perfect yet.
-
gif2gif: Quick and easy webui extension for dropping animated GIFs into img2img
Select the script, drop in a GIF, use img2img as normal to process it. Supports quick non-ffmpeg interpolation, and works surprisingly well with InstructPix2Pix. Intended to be a fun no-nonsense GIF pipeline.
-
NMKD Stable Diffusion GUI 1.9.0 is out now, featuring InstructPix2Pix - Edit images simply by using instructions! Link and details in comments.
Github Issue - Closed
What are some alternatives?
mmaction2 - OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark
stable-diffusion-webui - Stable Diffusion web UI
python-socketio - Python Socket.IO server and client
stable-diffusion-webui-instruct-pix2pix - Extension for webui to run instruct-pix2pix
react-native-sensors - A developer friendly approach for sensors in React Native
GFPGAN - GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.
conifer - Fast inference of Boosted Decision Trees in FPGAs
gif2gif - Automatic1111 Animated Image (input/output) Extension
conifer - Collect and revisit web pages.
k-diffusion - Karras et al. (2022) diffusion models for PyTorch
gsgen - [CVPR 2024] Text-to-3D using Gaussian Splatting
prolificdreamer - Official code of ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation (NeurIPS 2023 Spotlight)