Our great sponsors
-
cleanrl
High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
stable-baselines3
PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.
-
Deep-Reinforcement-Learning-Algorithms-with-PyTorch
PyTorch implementations of deep reinforcement learning algorithms and environments
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
machin
Reinforcement learning library(framework) designed for PyTorch, implements DQN, DDPG, A2C, PPO, SAC, MADDPG, A3C, APEX, IMPALA ...
cleanrl is the repo I base my code on, lots of single file implementations.
I borrow a lot of performance tricks from sample factory, which is awesome but hard to modify from its original APPO algorithm. rlpyt was more modular, and I borrowed more ideas from it (namedarraytuple), but still too limited.
I borrow a lot of performance tricks from sample factory, which is awesome but hard to modify from its original APPO algorithm. rlpyt was more modular, and I borrowed more ideas from it (namedarraytuple), but still too limited.
I tried tianshou and thought it was well-designed for modularity, but it was early in development when I tried and missing some basic features
stable baselines 3 is really starting to shape up, but I personally think it's better as a reference guide, there's a bit too much abstraction for the sake of ease of first plot, I wouldn't want to have to modify it.
MBRL-Lib for model-based RL
MTRL for multi-task RL
Machin is really nice, it is very easy to use and to try different things, although it’s developed by one person and maybe not appropriately tested yet.
Related posts
- [P] PettingZoo 1.24.0 has been released (including Stable-Baselines3 tutorials)
- SB3 - NotImplementedError: Box([-1. -1. -8.], [1. 1. 8.], (3,), <class 'numpy.float32'>) observation space is not supported
- [P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up
- Open RL Benchmark by CleanRL 0.5.0
- Show HN: An end-to-end reinforcement learning library for infinite horizon tasks