roboflow-100-benchmark
make-sense
roboflow-100-benchmark | make-sense | |
---|---|---|
8 | 7 | |
227 | 2,969 | |
4.0% | - | |
0.6 | 2.4 | |
6 months ago | about 2 months ago | |
Jupyter Notebook | TypeScript | |
MIT License | GNU General Public License v3.0 only |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
roboflow-100-benchmark
-
AI That Teaches Other AI
> Their SKILL tool involves a set of algorithms that make the process go much faster, they said, because the agents learn at the same time in parallel. Their research showed if 102 agents each learn one task and then share, the amount of time needed is reduced by a factor of 101.5 after accounting for the necessary communications and knowledge consolidation among agents.
This is a really interesting idea. It's like the reverse of knowledge distillation (which I've been thinking about a lot[1]) where you have one giant model that knows a lot about a lot & you use that model to train smaller, faster models that know a lot about a little.
Instead, you if you could train a lot of models that know a lot about a little (which is a lot less computationally intensive because the problem space is so confined) and combine them into a generalized model, that'd be hugely beneficial.
Unfortunately, after a bit of digging into the paper & Github repo[2], this doesn't seem to be what's happening at all.
> The code will learn 102 small and separte heads(either a linear head or a linear head with a task bias) for each tasks respectively in order. This step can be parallized on multiple GPUS with one task per GPU. The heads will be saved in the weight folder. After that, the code will learn a task mapper(Either using GMMC or Mahalanobis) to distinguish image task-wisely. Then, all images will be evaluated in the same time without a task label.
So the knowledge isn't being combined (and the agents aren't learning from each other) into a generalized model. They're just training a bunch of independent models for specific tasks & adding a model-selection step that maps an image to the most relevant "expert". My guess is you could do the same thing using CLIP vectors as the routing method to supervised models trained on specific datasets (we found that datasets largely live in distinct regions of CLIP-space[3]).
[1] https://github.com/autodistill/autodistill
[2] https://github.com/gyhandy/Shared-Knowledge-Lifelong-Learnin...
[3] https://www.rf100.org
- Roboflow 100: A New Object Detection Benchmark
-
[R] Roboflow 100: An open source object detection benchmark of 224,714 labeled images in novel domains to compare model performance
I'm Jacob, one of the authors of Roboflow 100, A Rich Multi-Domain Object Detection Benchmark, and I am excited to share our work with the community. In object detection, researchers are benchmarking their models on primarily COCO, and in many ways, it seems like a lot of these models are getting close to a saturation point. In practice, everyone is taking these models and finetuning them on their own custom dataset domains, which may vary from tagging swimming pools from Google Maps, to identifying defects in cell phones on an industrial line. We did some work to collect a representative benchmark of these custom domain problems by selecting from over 100,000 public projects on Roboflow Universe into 100 semantically diverse object detection datasets. Our benchmark comprises of 224,714 images, 11,170 labeling hourse, and 829 classes from the community for benchmarking on novel tasks. We also tried out the benchmark on a few popular models - comparing YOLOv5, YOLOv7, and the zero shot capabilities of GLIP. Use the benchmark here: https://github.com/roboflow-ai/roboflow-100-benchmark Paper link here: https://arxiv.org/pdf/2211.13523.pdf Or simply learn more here: https://www.rf100.org/ An immense thanks to the community, like this one, for making it possible to make this benchmark - we hope it moves the field forward! I'm around for any questions!
-
Introducing RF100: An open source object detection benchmark of 224,714 labeled images across 100 novel domains to compare model performance
Or simply learn more: https://www.rf100.org/
-
We took YOLOv5 and YOLOv7, trained them on 100 datasets, and compared their accuracy! š„ The results may surprise you.
github repository: https://github.com/roboflow-ai/roboflow-100-benchmark blogpost: https://blog.roboflow.com/roboflow-100/ arXiv paper: https://arxiv.org/abs/2211.13523
-
Show HN: Real-World Datasets for Benchmarking Object Detection Models
Github: https://github.com/roboflow-ai/roboflow-100-benchmark
At Roboflow, we've seen users fine-tune hundreds of thousands of computer vision models on custom datasets.
We observed that there's a huge disconnect between the types of tasks people are actually trying to perform in the wild and the types of datasets researchers are benchmarking their models on.
Datasets like MS COCO (with hundreds of thousands of images of common objects) are often used in research to compare models' performance, but then those models are used to find galaxies, look at microscope images, or detect manufacturing defects in the wild (often trained on small datasets containing only a few hundred examples). This leads to big discrepancies in models' stated and real-world performance.
We set out to tackle this problem by creating a new set of datasets that mirror many of the same types of challenges that models will face in the real world. We compiled 100 datasets from our community spanning a wide range of domains, subjects, and sizes.
We've benchmarked a couple of models (YOLOv5, YOLOv7, and GLIP) to start, but could use your help measuring the performance of others on this benchmark (check the GitHub for starter scripts showing how to pull the dataset, fine-tune models, and evaluate). We're very interested to learn which models do best in which real-world scenarios & to give researchers a new tool to make their models more useful for solving real-world problems.
make-sense
- Need help identifying a good open source data annotation tool
-
Free instance segmentation annotation tool
Hi šš»! Iām creator of https://makesense.ai. It supports Instance Segmentation. Take a look at the repo: https://github.com/SkalskiP/make-sense
-
Data Labelling Software
I created tool called MakeSense: https://github.com/SkalskiP/make-sense it is completely free and open sourced on GH
-
Roboflow 100: A New Object Detection Benchmark
Haven't heard of those two, but would be really awesome to see an integration. We have an open API[1] for just this reason: we really want to make it easy to use (and source) your data across all the different tools out there. We've recently launched integrations with other labeling[2] and AutoML[3] tools (and have integrations with the big-cloud AutoML tools as well[4]). We're hoping to have a bunch more integrations with other MLOps tools & platforms in 2023.
Re synthetic data specifically, we've written a couple of how-to guides for creating data from context augmentation[5], Unity Perception[6], and Stable Diffusion[7] & are talking to some others as well; it seems like a natural integration point (and someplace where we don't need to reinvent the wheel).
[1] https://docs.roboflow.com/rest-api
[2] https://github.com/SkalskiP/make-sense/pull/298
[3] https://github.com/ultralytics/yolov5/discussions/10425
[4] https://docs.roboflow.com/train/pro-third-party-training-int...
[5] https://blog.roboflow.com/how-to-create-a-synthetic-dataset-...
[6] https://blog.roboflow.com/unity-perception-synthetic-dataset...
[7] https://blog.roboflow.com/synthetic-data-with-stable-diffusi...
-
[Project] Football Players Tracking with YOLOv5 + ByteTRACK
Two things that carried me the most are my blog https://medium.com/@skalskip - which gave me my first job in computer vision, and my open-source GitHub project: https://github.com/SkalskiP/make-sense - which gave me all my jobs since I created it.
-
Hi everyone! I'm Piotr and for several years I have been developing a small open-source project for labeling photos - makesense.ai. I added a new feature this weekend. You can use YOLOv5 models to automatically annotate photos.
Link to GitHub project: https://github.com/SkalskiP/make-sense
-
Tool for human pose estimation keypoint annotation
I have also looked into make-sense and currently the docker and the npm refuse to work. I have already opened a ticket describing the issue .
What are some alternatives?
Shared-Knowledge-Lifelong-Learnin
label-studio - Label Studio is a multi-type data labeling and annotation tool with standardized output format
Shared-Knowledge-Lifelong-Learning - [TMLR] Lightweight Learner for Shared Knowledge Lifelong Learning
cvat - Annotate better with CVAT, the industry-leading data engine for machine learning. Used and trusted by teams at any scale, for data of any scale. [Moved to: https://github.com/cvat-ai/cvat]
roboflow-100-benchmark - Code for replicating Roboflow 100 benchmark results and programmatically downloading benchmark datasets [Moved to: https://github.com/roboflow/roboflow-100-benchmark]
AID - One-Stop System for Machine Learning.
fasterrcnn-pytorch-training-pipeline - PyTorch Faster R-CNN Object Detection on Custom Dataset
VoTT - Visual Object Tagging Tool: An electron app for building end to end Object Detection Models from Images and Videos.
autodistill - Images to inference with no labeling (use foundation models to train supervised models).
Universal Data Tool - Collaborate & label any type of data, images, text, or documents, in an easy web interface or desktop app.
yolov5 - YOLOv5 š in PyTorch > ONNX > CoreML > TFLite
SynthDet - SynthDet - An end-to-end object detection pipeline using synthetic data