Track-Anything
lama
Track-Anything | lama | |
---|---|---|
16 | 17 | |
6,104 | 7,227 | |
- | 2.5% | |
8.1 | 5.4 | |
3 months ago | 3 months ago | |
Python | Jupyter Notebook | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Track-Anything
-
Keying/masking person on a footage
I was doing rotoscoping for a silhouette of a girl dancing in front of a building, then I saw this amazing tool: https://github.com/gaomingqi/Track-Anything
-
Advice for multi-animal tracking for scientific research?
My question is, how can we modernize this pipeline? We've experimented a bit with the new SAM-based track-anything tool, and it seems promising, but we actually don't want to "track anything", we only want to track fishes. What would you do in 2023, to extract tracks of one specific class of object from long video datasets? I'm hoping for any advice at all.
-
[D] Which open source models can replicate wonder dynamics's drag'n'drop cg characters?
The Track-Anything tool already implements this
- Tutorial for Track-Anything, an interactive tool to segment, track, and inpaint anything in videos.
- Github for Track Anything
-
Segment Anything for Video - Track Anything! 🤖
Segment Anything for Video - Track Anything! 🤖 With this tool, you can automatically isolate objects, make edits using inpainting, and track objects with precision. It's a game-changer for creative projects. Even though it does not work well with the shadows yet, we expect a rapid evolution of these technologies. Github : https://github.com/gaomingqi/Track-Anything)
-
How to adapt an existing Python project to my specific use case without bringing in unnecessary dependencies or reinventing the wheel
I would like to know if there are any guidelines to follow when adapting an existing Python project for my own use-case. Specifically, I want to customize the output of the Track-Anything project by incorporating my own processing steps. However, I do not want to import the entire codebase. Rather, I only want to import the minimum amount of code necessary to produce the same output with object tracking, without having to reimplement functions that are already available.
- Track-Anything should get implemented in kdelive
-
SUSTech VIP Lab Proposes Track Anything Model (TAM) That Achieves High-Performance Interactive Tracking and Segmentation in Videos
Here is the GitHub: https://github.com/gaomingqi/track-anything
lama
-
Can someone please help me with inpainting settings to remove the subject from this image? I want to rebuild as much of the original background as possible.
You could try to use ControlNet inpaint+lama locally, but results aren't as good in my experience. Or you could try local install of lama directly, but the setup process isn't very smooth.
-
[D] Which open source models can replicate wonder dynamics's drag'n'drop cg characters?
You may be able to remove the actor with lama. https://github.com/advimman/lama
-
ControlNet Update: [1.1.222] Preprocessor: inpaint_only+lama
LaMa: Resolution-robust Large Mask Inpainting with Fourier Convolutions (Apache-2.0 license) Roman Suvorov, Elizaveta Logacheva, Anton Mashikhin, Anastasia Remizova, Arsenii Ashukha, Aleksei Silvestrov, Naejin Kong, Harshith Goka, Kiwoong Park, Victor Lempitsky (Samsung Research and EPFL)
-
[Inpainting] [Q] Want to remove a person/ group of people from an image.
You could try using LaMA: https://github.com/saic-mdal/lama .
-
[task] Python developer for two tasks, one 30$ the other 50$
Task 1: 30$ Write a script to prepare data for https://github.com/saic-mdal/lama. You should be able to prove on video that your script works properly after i give you some data and you train and there's some acceptable output (not perfect, since i know how ml works).
-
Research Topics for Master/PHD
I have some interest in image inpainting, but the recent paper introduced by Samsung - https://github.com/saic-mdal/lama seems like it reached a stage where there isn't much room that we can add for novelty. I am quite interested but not confident if I can come up with a novel idea that can do better than the proposed model, especially if they trained the model on multiple GPUs for days, which is something that I don't have.
- The Black Hole Photographs: Censored Images from America’s Great Depression
- Image inpainting tool powered by LaMa
- Resolution-Robust Large Mask Inpainting with Fourier Convolutions
-
[D] Paper Explained - Resolution-robust Large Mask Inpainting with Fourier Convolutions (w/ Author Interview)
Code: https://github.com/saic-mdal/lama
What are some alternatives?
stable-diffusion-webui - Stable Diffusion web UI
IOPaint - Image inpainting tool powered by SOTA AI Model. Remove any unwanted object, defect, people from your pictures or erase and replace(powered by stable diffusion) any thing on your pictures.
segment-anything - The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Torrent-To-Google-Drive-Downloader-v3 - Simple notebook to stream torrent files to Google Drive using Google Colab and python3.
sam-clip - Use Grounding DINO, Segment Anything, and CLIP to label objects in images.
strv-ml-mask2face - Virtually remove a face mask to see what a person looks like underneath
XMem - [ECCV 2022] XMem: Long-Term Video Object Segmentation with an Atkinson-Shiffrin Memory Model
cleanup.pictures - Code for https://cleanup.pictures
sd-webui-segment-anything - Segment Anything for Stable Diffusion WebUI
Colab-Crypto-Mining - Cryptocurrency Mining Experiments on Google CoLab Notebooks
dl-colab-notebooks - Try out deep learning models online on Google Colab
Real-time-Object-Detection-for-Autonomous-Driving-using-Deep-Learning - My Computer Vision project from my Computer Vision Course (Fall 2020) at Goethe University Frankfurt, Germany. Performance comparison between state-of-the-art Object Detection algorithms YOLO and Faster R-CNN based on the Berkeley DeepDrive (BDD100K) Dataset.