The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning. Learn more →
Top 23 Python segment-anything Projects
-
Track-Anything
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
InternGPT
InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)
-
segment-geospatial
A Python package for segmenting geospatial data with the Segment Anything Model (SAM)
-
anylabeling
Effortless AI-assisted data labeling with AI support from YOLO, Segment Anything, MobileSAM!!
-
Caption-Anything
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
autodistill
Images to inference with no labeling (use foundation models to train supervised models).
-
multimodal-maestro
Effective prompting for Large Multimodal Models like GPT-4 Vision, LLaVA or CogVLM. 🔥
-
sd-webui-inpaint-anything
Inpaint Anything extension performs stable diffusion inpainting on a browser UI using masks from Segment Anything.
-
GLEE
[CVPR2024 Highlight]GLEE: General Object Foundation Model for Images and Videos at Scale (by FoundationVision)
-
awesome-foundation-and-multimodal-models
👁️ + 💬 + 🎧 = 🤖 Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]
-
comfyui_segment_anything
Based on GroundingDino and SAM, use semantic strings to segment any element in an image. The comfyui version of sd-webui-segment-anything.
-
TinySAM
Official PyTorch implementation of "TinySAM: Pushing the Envelope for Efficient Segment Anything Model"
-
Instruct2Act
Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
-
inpaint-anything
Inpaint Anything performs stable diffusion inpainting on a browser UI using masks from Segment Anything. (by Uminosachi)
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
I was doing rotoscoping for a silhouette of a girl dancing in front of a building, then I saw this amazing tool: https://github.com/gaomingqi/Track-Anything
Segment Anything in High Quality (https://arxiv.org/abs/2306.01567)
Project mention: Textual inversion. The best way to prepare photos of a person? | /r/StableDiffusion | 2023-12-06One idea would be to use Segment Anything to cut out the character/face from the background and then replace with random backgrounds that you generate with stable diffusion. Here's an extension for Automatic1111 :) https://github.com/continue-revolution/sd-webui-segment-anything
You can also create an issue and ask the developers for help.
Maybe try reporting as a GitHub issue? https://github.com/opengeos/segment-geospatial/issues
Project mention: AnyLabeling Auto-labeling with MobileSAM - the newest and fastest variant of Segment Anything | /r/computervision | 2023-06-28Check out AnyLabeling v0.3.2 today: https://github.com/vietanhdev/anylabeling/releases/tag/v0.3.2.
Roboflow | Open Source Software Engineer, Web Designer / Developer, and more. | Full-time (Remote, SF, NYC) | https://roboflow.com/careers?ref=whoishiring0224
Roboflow is the fastest way to use computer vision in production. We help developers give their software the sense of sight. Our end-to-end platform[1] provides tooling for image collection, annotation, dataset exploration and curation, training, and deployment.
Over 250k engineers (including engineers from 2/3 Fortune 100 companies) build with Roboflow. We now host the largest collection of open source computer vision datasets and pre-trained models[2]. We are pushing forward the CV ecosystem with open source projects like Autodistill[3] and Supervision[4]. And we've built one of the most comprehensive resources for software engineers to learn to use computer vision with our popular blog[5] and YouTube channel[6].
We have several openings available but are primarily looking for strong technical generalists who want to help us democratize computer vision and like to wear many hats and have an outsized impact. Our engineering culture is built on a foundation of autonomy & we don't consider an engineer fully ramped until they can "choose their own loss function". At Roboflow, engineers aren't just responsible for building things but also for helping us figure out what we should build next. We're builders & problem solvers; not just coders. (For this reason we also especially love hiring past and future founders.)
We're currently hiring full-stack engineers for our ML and web platform teams, a web developer to bridge our product and marketing teams, several technical roles on the sales & field engineering teams, and our first applied machine learning researcher to help push forward the state of the art in computer vision.
[1]: https://roboflow.com/?ref=whoishiring0224
[2]: https://roboflow.com/universe?ref=whoishiring0224
[3]: https://github.com/autodistill/autodistill
[4]: https://github.com/roboflow/supervision
[5]: https://blog.roboflow.com/?ref=whoishiring0224
[6]: https://www.youtube.com/@Roboflow
Project mention: Show HN: Multimodal Maestro – Prompt tools for use with LMMs | news.ycombinator.com | 2023-11-29
Hey everyone,Today I installed inpaint-anything extension from github(https://github.com/Uminosachi/sd-webui-inpaint-anything) to Stable Diffusion. It works perfectly.
Project mention: Sharing an automatic tool I built for video background removal (free and better than runwayml) | /r/StableDiffusion | 2023-05-04
Part of this process can be a remastering step also for old videos. Masking backgrounds across frames and performing super-resolution with all known references and scaling details for characters using fine-tuned models for each actor. We have a lot new SAM tools to assist with this process. It probably won't be magically done for a while, but a few people could remaster a show rather than a large team.
Project mention: When SAM Meets NeRF: This AI Model Can Segment Anything in 3D | /r/machinelearningnews | 2023-05-22
Following the comments to this old post, I tried to use in-painting with manual mask selection. I didn't get beautiful results but I'm sure with some tweaking, I could make it better. The main problem I had was having to manually select the area where I wanted to place the logo and trying to resize my logo mask to the fit the segment. I tried some automatic segmentation tools (Clipseg and Segment Anything). I couldn't tell the segmentation models to find a good area to for logo placement (i.e. some small flat surface). Given the complexity of what I was dealing with, I think there could be a better way (XY problem).
Project mention: [R]Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model | /r/MachineLearning | 2023-05-20Code: https://github.com/OpenGVLab/Instruct2Act
Project mention: How to change / custom Model ID of HuggingFace Model | /r/StableDiffusion | 2023-10-15I am using Inpaint-Anything. In that they are using these & these type of models for inpainting, I think they called diffusers, sorry if I am wrong idk couz I am new to this and I want to create absolutereality_v181INPAINTING.safetensors model for inpainting like above type of models.
Project mention: Hi-Sam: Marrying Segment Anything Model for Hierarchical Text Segmentation | news.ycombinator.com | 2024-02-21
fiftyone plugins download https://github.com/jacobmarks/zero-shot-prediction-plugin
Project mention: [P] Made a simple semantic segmentation annotation tool with segment-anything masks support in PyQt5 | /r/MachineLearning | 2023-09-26I just open-sourced (MIT License) semantic segmentation annotation tool powered by segment-anything model that I used for a while in my projects. Hopefully it will help someone as it seems to me that it is more suitable for small projects than popular huge web based annotation tools. Link to the project: SAMAT (any feedback in Discussions section on GitHub is appreciated) Features:
Project mention: SAM-CLIP: Combine Grounded Sam and Clip for Instance Segmentation | news.ycombinator.com | 2023-11-17
Python segment-anything related posts
- How to blend a logo or clip art to a design
- Textual inversion. The best way to prepare photos of a person?
- Keying/masking person on a footage
- AnyLabeling Auto-labeling with MobileSAM - the newest and fastest variant of Segment Anything
- Best way to mask images automatically?
- Advice for multi-animal tracking for scientific research?
- Segment Anything in High Quality
-
A note from our sponsor - WorkOS
workos.com | 27 Apr 2024
Index
What are some of the best open-source segment-anything projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | Track-Anything | 6,090 |
2 | sam-hq | 3,347 |
3 | sd-webui-segment-anything | 3,192 |
4 | InternGPT | 3,121 |
5 | segment-geospatial | 2,652 |
6 | anylabeling | 1,852 |
7 | Caption-Anything | 1,600 |
8 | autodistill | 1,529 |
9 | multimodal-maestro | 942 |
10 | sd-webui-inpaint-anything | 914 |
11 | segment-anything-video | 907 |
12 | sam-pt | 905 |
13 | GLEE | 879 |
14 | SegmentAnythingin3D | 785 |
15 | awesome-foundation-and-multimodal-models | 509 |
16 | comfyui_segment_anything | 384 |
17 | TinySAM | 356 |
18 | Instruct2Act | 254 |
19 | inpaint-anything | 154 |
20 | Hi-SAM | 120 |
21 | zero-shot-prediction-plugin | 23 |
22 | samat | 23 |
23 | sam-clip | 20 |
Sponsored