Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
tflite2tensorflow
Generate saved_model, tfjs, tf-trt, EdgeTPU, CoreML, quantized tflite, ONNX, OpenVINO, Myriad Inference Engine blob and .pb from .tflite. Support for building environments with Docker. It is possible to directly access the host PC GUI and the camera to verify the operation. NVIDIA GPU (dGPU) support. Intel iHD GPU (iGPU) support. Supports inverse quantization of INT8 quantization model.
ToucanTrack can be found at the following repository: https://github.com/noahcoolboy/toucan-track/tree/main
They come in the form of tflite models, so I had to convert them to onnx. I used tf2onnx for converting the pose landmark model and tflite2tensorflow for converting the pose detection model. For improving performance, I had created a small script which modified the landmark models for supporting batch inference. This script is not included in the repository, but do tell me if you need it!
They come in the form of tflite models, so I had to convert them to onnx. I used tf2onnx for converting the pose landmark model and tflite2tensorflow for converting the pose detection model. For improving performance, I had created a small script which modified the landmark models for supporting batch inference. This script is not included in the repository, but do tell me if you need it!
If you're looking for the differences in terms of how inference is done, I recommend you take a look at MediaPipe's source code. MediaPipe doesn't use raw code, but uses a "graph" instead (eg. pose_landmark_cpu.pbtxt), which can be visualised using MediaPipe Viz. I also used axinc-ai/ailia-models as the base (preprocessing, inference, postprocessing, etc...) which I further built upon (keypoint refinement, roi from keypoints, filtering / smoothing, etc...)
If you're looking for the differences in terms of how inference is done, I recommend you take a look at MediaPipe's source code. MediaPipe doesn't use raw code, but uses a "graph" instead (eg. pose_landmark_cpu.pbtxt), which can be visualised using MediaPipe Viz. I also used axinc-ai/ailia-models as the base (preprocessing, inference, postprocessing, etc...) which I further built upon (keypoint refinement, roi from keypoints, filtering / smoothing, etc...)
This looks amazing! I really would like to try it out but I don't feel like downloading CL-Eye Platform SDK from a dropbox. Especially since this is usually a paid product. Are there any possible alternatives for this driver? Something like this one? Thank you!