rex-gym
gretel-synthetics
Our great sponsors
rex-gym | gretel-synthetics | |
---|---|---|
1 | 4 | |
957 | 530 | |
- | 4.3% | |
0.0 | 7.3 | |
about 1 year ago | 8 days ago | |
Python | Python | |
Apache License 2.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
rex-gym
-
By moving the battery pack forward, you can make the popular SpotMicro design balance much better. We had trouble getting it to do standing/walking because the center of mass was far to the back.
Our work was based on (SpotMicro)[https://github.com/michaelkubina/SpotMicroESP32] and (Rex Gym)[https://github.com/nicrusso7/rex-gym]. Our GitHub is (here)[https://github.com/LSaldyt/laser-dog]
gretel-synthetics
-
Ask HN: If we train an LLM with “data” instead of “language” tokens
Hey there! Co-founder of Gretel.ai here, and I think I can provide some insights on this topic.
Firstly, the concept you're hinting at is not purely traditional ML. In traditional machine learning, we often prioritize feature extraction and engineering specific to a given problem space before training.
What you're describing and what we've been working on at Gretel.ai, is leveraging the power of models like Large Language Models (LLMs) to understand and extrapolate from vast amounts of diverse data without the need for time-consuming feature engineering. Here's a link to our open-source library https://github.com/gretelai/gretel-synthetics for synthetic data generation (currently supporting GAN and RNN-based language models), and also our recent announcement around a Tabular LLM we're training to help people build with data https://gretel.ai/tabular-llm
A few areas where we've found tabular or Large Data Models to be really useful are:
-
Libraries for synthetic data?
you can try QuantGAN: https://github.com/PakAndrey/QuantGANforRisk also try DoppelGANger https://github.com/gretelai/gretel-synthetics/tree/master/src/gretel_synthetics/timeseries_dgan
- Which open source tool for generating synthetic data sets?
- Gretel-synthetics: open-source library to create synthetic datasets
What are some alternatives?
pybullet-gym - Open-source implementations of OpenAI Gym MuJoCo environments for use with the OpenAI Gym Reinforcement Learning Research Platform.
Copulas - A library to model multivariate data using copulas.
robot-gym - RL applied to robotics.
gretel-python-client - The Gretel Python Client allows you to interact with the Gretel REST API.
drl_grasping - Deep Reinforcement Learning for Robotic Grasping from Octrees
adversarial-robustness-toolbox - Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
stable-baselines3 - PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.
CTGAN - Conditional GAN for generating synthetic tabular data.
PILCO - Bayesian Reinforcement Learning in Tensorflow
RobustVideoMatting - Robust Video Matting in PyTorch, TensorFlow, TensorFlow.js, ONNX, CoreML!
gym-battleship - Battleship environment for reinforcement learning tasks
AI-basketball-analysis - :basketball::robot::basketball: AI web app and API to analyze basketball shots and shooting pose.