temporal-shift-module
CushyStudio
Our great sponsors
temporal-shift-module | CushyStudio | |
---|---|---|
3 | 2 | |
2,019 | 571 | |
0.9% | - | |
3.0 | 10.0 | |
7 months ago | 10 days ago | |
Python | TypeScript | |
MIT License | GNU Affero General Public License v3.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
temporal-shift-module
- Stable Video Diffusion
-
Can two-stream networks trained for video action recognition be used for real-time usecases?
My question mostly has to do with optical flow. One of the two-stream networks I'm interested in trying out is TSN-TSM, as there are pre-trained weights available for it on the Assembly101 dataset released a few months ago.
-
I am having a hard time understanding this paper(Temporal shift module). Can some who have read it before or willing to read it explain me better in a more elaborate way?
This is the paper. (https://arxiv.org/abs/1811.08383). Here they are talking about how they can achieve temporal modelling by moving channels, which I assume are the RGB channels across frames. But I am super confused by the lingo. Here is the repo (https://github.com/mit-han-lab/temporal-shift-module). I can't give better rewards except virtual hugs. Thank you.
CushyStudio
-
Stable Video Diffusion
Finally ! Now that this is out, I can finally start adding proper video widgets to CushyStudio https://github.com/rvion/CushyStudio#readme . Really hope I can get in touch with StabilityAi people soon. Maybe Hacker News will help
-
How to Build Your Own AI-Generated Images with ControlNet and Stable Diffusion
I'm building CushyStudio https://github.com/rvion/cushystudio#readme to make Stable Diffusion practical and fun to play with.
It's still a bit rough around the corners, and I haven't properly launched it yet, but if you want to play with ControlNets, pre-processors, IP adapters, and all those various SD technologies, it's a pretty fun tool ! I personally use for real-time scribble to image, things like this :)
(will post that properly on HN in a few days / week I think, when early feedback will have been properly addressed)
What are some alternatives?
mmaction2 - OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark
Fooocus - Focus on prompting and generating
python-socketio - Python Socket.IO server and client
Stable-Diffusion - Stable Diffusion, SDXL, LoRA Training, DreamBooth Training, Automatic1111 Web UI, DeepFake, Deep Fakes, TTS, Animation, Text To Video, Tutorials, Guides, Lectures, Courses, ComfyUI, Google Colab, RunPod, NoteBooks, ControlNet, TTS, Voice Cloning, AI, AI News, ML, ML News, News, Tech, Tech News, Kohya LoRA, Kandinsky 2, DeepFloyd IF, Midjourney
react-native-sensors - A developer friendly approach for sensors in React Native
p5.capture - 🎬 super easy recording for p5.js animations
conifer - Fast inference of Boosted Decision Trees in FPGAs
LLM-groundedDiffusion - LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models (LLM-grounded Diffusion: LMD)
conifer - Collect and revisit web pages.
react - A wrapper component that allows you to utilise P5 sketches within React apps.
gsgen - [CVPR 2024] Text-to-3D using Gaussian Splatting
pts - A library for visualization and creative-coding