panda-gym
dreamerv2
panda-gym | dreamerv2 | |
---|---|---|
3 | 4 | |
450 | 853 | |
- | - | |
5.3 | 0.0 | |
5 months ago | over 1 year ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
panda-gym
-
Hyperparameters for pick&place with Franka Emika manipulator
I'm trying to solve pick&place (and possibly also the other tasks in this repository) with Franka Emika Panda manipulator implemented in Mujoco. I've tried for long with stable_baseline3 but without any results, someone told me to try with RLLib because has better implementation (?), but still I can't find any solution...
-
SAFE-PANDA-GYM a modification to Panda - Gym to train Safe-RL agents
We develop a modification to the Panda Gym by adding constraints to the environments like Unsafe regions and, constraints on the task. The aim is to develop an environment to test CMDPs (Constraint Markov Decision Process) / Safe-RL algorithms such as CPO, PPO - Lagrangian and algorithms developed by the team. Agents would not only have to come up with optimal policy for control and planning but also ensure they don't violate a constraint.
-
Did anyone try Panda-Gym?
The acquisition of Mujoco makes Openai to remove the robotics from their repo. I had no choice but to find an alternative. Then I found https://github.com/qgallouedec/panda-gym which is built on PyBullet.
dreamerv2
-
Sources of Actor Gradients
In fact, they found that just reinforce gradients work in DM control now too: Dreamerv2 GitHub (they just needed to turn off gradients through the action path - which I guess was being passed back with straight-through estimation? I'm actually having a difficult time telling how the gradient is different on the action vs policy.log_prob(action)).
-
PyDreamer: model-based RL written in PyTorch + integrations with DM Lab and MineRL environments
This is my implementation of Hafner et al. DreamerV2 algorithm. I found the PlaNet/Dreamer/DreamerV2 paper series to be some of the coolest RL research in recent years, showing convincingly that MBRL (model-based RL) does work and is competitive with model-free algorithms. And we all know that AGI will be model-based, right? :)
-
Any current state or the art libraries for training agents to play atari games?
Last I checked, for running off a single node, the state of the art was Dreamerv2 https://github.com/danijar/dreamerv2
- Google AI, DeepMind And The University of Toronto Introduce DreamerV2, The First Reinforcement Learning (RL) Agent That Outperforms Humans on The Atari Benchmark
What are some alternatives?
dm_env - A Python interface for reinforcement learning environments
dreamerv3 - Mastering Diverse Domains through World Models
stable-baselines3 - PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.
dreamer - Dream to Control: Learning Behaviors by Latent Imagination
sapai - Super auto pets engine built with reinforment learning training in mind
dm_control - Google DeepMind's software stack for physics-based simulation and Reinforcement Learning environments, using MuJoCo.
habitat-api - A modular high-level library to train embodied AI agents across a variety of tasks, environments, and simulators. [Moved to: https://github.com/facebookresearch/habitat-lab]
stable-baselines3-contrib - Contrib package for Stable-Baselines3 - Experimental reinforcement learning (RL) code
planet - Learning Latent Dynamics for Planning from Pixels
Safe-panda-gym - OpenaAI Gym Franka Emika Panda robot environment based on PyBullet.
orion - Asynchronous Distributed Hyperparameter Optimization.