How to improve a YoloV5 model after the first training?

This page summarizes the projects mentioned and recommended in the original post on /r/computervision

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • Yolo_mark

    GUI for marking bounded boxes of objects in images for training neural network Yolo v3 and v2

  • My work heavily involves the use of the yolo algorithm such as optimising it for performance on mobile devices. Yolov5 is made by a private company that has been pushing sub bar models for a while. I've benchmarked their smallest models comparing them to Yolov4 tiny and the results were staggering, v4 being around 3-4 times faster. Yolov4 has way more resources for development, I highly suggest checking out this repo https://github.com/AlexeyAB/darknet

  • DarkLabel

    Video/Image Labeling and Annotation Tool

  • There is still no point retraining your model with its own output tho. You will get better results annotating it yourself. You can make use of this annotation tool https://github.com/darkpgmr/DarkLabel to save annotation time. If your training images are frames in a video, by using this tool you can propagate your bounding boxes into future frames. The boxes will eventually drift tho, so you will have to reinitialise the boxes every now and then.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • sort

    Simple, online, and realtime tracking of multiple objects in a video sequence.

  • One interesting thing to consider is about using a new model's output as training data is the following. If you use some object tracking algorithm like SORT, you might be able to find frames where the object was "tracked" even though the model didn't detect the object (some type of flickering). In this case, you're able to run the model on video and make it better by giving it these new bounding boxes based on "tracking". Does that make sense?

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Apple to Power AI Features with M2 Ultra Servers

    2 projects | news.ycombinator.com | 10 May 2024
  • Show HN: An SQS Alternative on Postgres

    10 projects | news.ycombinator.com | 9 May 2024
  • Show HN: OpenVoice_server, a simple API server built on top of OpenVoice

    1 project | news.ycombinator.com | 10 May 2024
  • This Week In Python

    5 projects | dev.to | 10 May 2024
  • PyTorch 2.3: User-Defined Triton Kernels, Tensor Parallelism in Distributed

    1 project | news.ycombinator.com | 10 May 2024