Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
MotionBERT
[ICCV 2023] PyTorch Implementation of "MotionBERT: A Unified Perspective on Learning Human Motion Representations"
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
Summary: I've implemented real-time mono-webcam full body tracking, but I'm having trouble sending this data to vrchat. This function was actually implemented on Unity, and socket communication was used in Unity. Is there a socket communication method that works with udonsharp or vrchat? Here are the solutions I reviewed. https://github.com/Xytabich/UNet I think Unet is very similar to the solution I was looking for, but the connection ID uses the person object that entered the world. I want to receive data from outside.
https://github.com/fa-m/midi-websocket MIDI websockets pose a fundamental question to me. This solution is difficult to analyze and emphasizes the importance of security.
This only works for your own avatar though. If you want to make something that is displayed in a single world you can look into ShaderMotion, which is a system to bring humanoid poses into VRC worlds. The data is sent via a video feed (as blinking pixels) and a shader then reads out those pixels and "animates" the model accordingly.
Thank you for answer. The method I am using is one of the monocular 3d pose estimation studies, and is as follows. https://github.com/Walter0807/MotionBERT
Lox also made OSCMotion, which allows you to send the joint positions via OSC instead. This is pretty scuffed though, since the update rate is far too low for this to look good. It is possible in theory though.