b3d_mocap_import
mediapipe
b3d_mocap_import | mediapipe | |
---|---|---|
1 | 49 | |
106 | 25,610 | |
- | 1.6% | |
1.8 | 9.9 | |
over 2 years ago | 4 days ago | |
Python | C++ | |
- | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
b3d_mocap_import
-
My experience using VIBE ( free motion extraction tool for animation )
Regarding installation, I didn't do step 14 and 21 and it worked for me. I also tried to replace steps 17-24 with this addon. It did import VIBE's pickles directly, but they are not usual pickles, so they require to install external dependency joblib to the Blender's bundled Python. And the data inside are numpy objects, so it is not exactly an exchange format you would want to use. However, the addon was unable to import the animations correctly in 2.92, so I gave up on this idea. So I end up registering and downloading the models, then dumped the full Blender's distribution into the venv folder and it worked.
mediapipe
-
Mediapipe openpose Controlnet model for SD
mediapipe/docs/solutions/pose.md at master · google/mediapipe · GitHub
-
MEDIAPIPE on-device diffusion plugins for conditioned text-to-image generation
Today, we announce MediaPipe diffusion plugins, which enable controllable text-to-image generation to be run on-device. Expanding upon our prior work on GPU inference for on-device large generative models, we introduce new low-cost solutions for controllable text-to-image generation that can be plugged into existing diffusion models and their Low-Rank Adaptation (LoRA) variants.
-
Running a TensorFlow object detector model and drawing boxes around objects at 60 FPS - all in React Native / JavaScript!
You can just grab the TFLite version! https://github.com/google/mediapipe/blob/master/docs/solutions/models.md
-
OpenAI came after our domain because we use GPT in it
I believe Google already released transformers under an apache 2 license with a patent grant:
https://github.com/google/mediapipe/blob/master/mediapipe/mo...
-
Open source Background Remover: Remove Background from images and video using AI
I was going to say that I like the MediaPipe Selfie Segmentation model for doing this sort of thing in a web page, but I've just noticed (when getting the GitHub link[1]) that Google have marked the code as legacy[2] ... no idea if the new solution is better/easier to use[3].
For what it's worth, my CodePen using the old model is here: https://codepen.io/kaliedarik/pen/PopBxBM
[1] - https://github.com/google/mediapipe/blob/master/docs/solutio...
[2] - "Attention: Thank you for your interest in MediaPipe Solutions. As of April 4, 2023, this solution was upgraded to a new MediaPipe Solution."
[3] - https://developers.google.com/mediapipe/solutions/vision/ima...
-
[P] Pattern recognition
I have used mediapipe very successfully in multiple projects and it's very easy to get running. You can choose from many different vision tasks including hand landmarks ( https://github.com/google/mediapipe/blob/master/docs/solutions/hands.md )
-
Getting face feature pose statistics
I found MediaPipe's Face Mesh and was impressed with how simple it was to get going, but it just gives you the landmark points and I've not gone any further yet.
-
New ControlNet Face Model
We've trained ControlNet on a subset of the LAION-Face dataset using modified output from MediaPipe's face mesh annotator to provide a new level of control when generating images of faces.
-
Trained an ML model using TensorFlow.js to classify American Sign Language (ASL) alphabets on browser. We are creating an open-source platform and would love to receive your feedback on our project.
Medipaipe library link: https://mediapipe.dev/
-
mediapipe VS daisykit - a user suggested alternative
2 projects | 24 Mar 2023
What are some alternatives?
frankmocap - A Strong and Easy-to-use Single View 3D Hand+Body Pose Estimator
openpose - OpenPose: Real-time multi-person keypoint detection library for body, face, hands, and foot estimation
VIBE - Official implementation of CVPR2020 paper "VIBE: Video Inference for Human Body Pose and Shape Estimation"
ue4-mediapipe-plugin - UE4 MediaPipe plugin
vibe_win_install
Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration
AlphaPose - Real-Time and Accurate Full-Body Multi-Person Pose Estimation&Tracking System
BlazePose-tensorflow - A third-party Tensorflow Implementation for paper "BlazePose: On-device Real-time Body Pose tracking".
jeelizFaceFilter - Javascript/WebGL lightweight face tracking library designed for augmented reality webcam filters. Features : multiple faces detection, rotation, mouth opening. Various integration examples are provided (Three.js, Babylon.js, FaceSwap, Canvas2D, CSS3D...).
pifuhd - High-Resolution 3D Human Digitization from A Single Image.
flutter_hand_tracking_plugin - 这是一个 Flutter Packge 以实现摄像头精确追踪并识别十指的运动路径/轨迹和手势动作, 且输出22个手部关键点以支持更多手势自定义. 基于这个包可以编写业务逻辑将手势信息实时转化为指令信息: 一二三四五, rock, spiderman...还可以对不同手势编写不同特效. 可用于短视频直播特效, 智能硬件等领域, 为人机互动带来更自然丰富的体验
tfjs-models - Pretrained models for TensorFlow.js