Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Stable-baselines3 Alternatives
Similar projects and alternatives to stable-baselines3
-
Ray
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
-
cleanrl
High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
stable-baselines
A fork of OpenAI Baselines, implementations of reinforcement learning algorithms
-
-
Pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
-
PettingZoo
An API standard for multi-agent reinforcement learning environments, with popular reference environments and related utilities
-
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
Super-mario-bros-PPO-pytorch
Proximal Policy Optimization (PPO) algorithm for Super Mario Bros
-
rl-baselines3-zoo
A training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included.
-
SuperSuit
A collection of wrappers for Gymnasium and PettingZoo environments (being merged into gymnasium.wrappers and pettingzoo.wrappers
-
stable-baselines3-contrib
Contrib package for Stable-Baselines3 - Experimental reinforcement learning (RL) code
-
gym
A toolkit for developing and comparing reinforcement learning algorithms.
-
baselines
OpenAI Baselines: high-quality implementations of reinforcement learning algorithms
-
Deep-Reinforcement-Learning-Algorithms-with-PyTorch
PyTorch implementations of deep reinforcement learning algorithms and environments
-
pytorch-trpo
PyTorch implementation of Trust Region Policy Optimization
-
machin
Reinforcement learning library(framework) designed for PyTorch, implements DQN, DDPG, A2C, PPO, SAC, MADDPG, A3C, APEX, IMPALA ...
-
Tic-Tac-Toe-Gym
This is the Tic-Tac-Toe game made with Python using the PyGame library and the Gym library to implement the AI with Reinforcement Learning
-
-
RL-Adventure
Pytorch Implementation of DQN / DDQN / Prioritized replay/ noisy networks/ distributional values/ Rainbow/ hierarchical RL
-
agents
TF-Agents: A reliable, scalable and easy to use TensorFlow library for Contextual Bandits and Reinforcement Learning.
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
stable-baselines3 reviews and mentions
-
Sim-to-real RL pipeline for open-source wheeled bipeds
The latest release (v3.0.0) of Upkie's software brings a functional sim-to-real reinforcement learning pipeline based on Stable Baselines3, with standard sim-to-real tricks. The pipeline trains on the Gymnasium environments distributed in upkie.envs (setup: pip install upkie) and is implemented in the PPO balancer. Here is a policy running on an Upkie:
-
[P] PettingZoo 1.24.0 has been released (including Stable-Baselines3 tutorials)
PettingZoo 1.24.0 is now live! This release includes Python 3.11 support, updated Chess and Hanabi environment versions, and many bugfixes, documentation updates and testing expansions. We are also very excited to announce 3 tutorials using Stable-Baselines3, and a full training script using CleanRL with TensorBoard and WandB.
-
SB3 - NotImplementedError: Box([-1. -1. -8.], [1. 1. 8.], (3,), <class 'numpy.float32'>) observation space is not supported
Therefore, I debugged this error to the ReplayBuffer that was imported from `SB3`. This is the problem function -
-
Shimmy 1.0: Gymnasium & PettingZoo bindings for popular external RL environments
Have you ever wanted to use dm-control with stable-baselines3? Within Reinforcement learning (RL), a number of APIs are used to implement environments, with limited ability to convert between them. This makes training agents across different APIs highly difficult, and has resulted in a fractured ecosystem.
-
Stable-Baselines3 v1.8 Release
Changelog: https://github.com/DLR-RM/stable-baselines3/releases/tag/v1.8.0
-
[P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up
Great project! One question though, is there any reason why you are not using existing RL models instead of creating your own, such as stable baselines?
-
Is Stable Baselines 3 no longer compatible with PettingZoo?
I was able to get Stable Baselines 3 to work with gymnasium by following the details in this work-in-progress PR: https://github.com/DLR-RM/stable-baselines3/pull/780. I have not used PettingZoo, though.
- [OC] Inteligência Artificial jogando Mega Man!
-
New to reinforcement learning.
I'd say this is a great path but I'd also look at the basic on-policy gradient actor critic methods like A2C and eventually PPO. Someone recommended SAC which also really good. There are tons of environments in the https://github.com/Farama-Foundation/PettingZoo as well if you want to mess with those. You can also check out stable baselines https://github.com/DLR-RM/stable-baselines3 which is pretty popular. If you want to get into the theory more I recommend reading the Sutton and Barto book on reinforcement learning.
-
How to proceed further? (Learning RL)
If you want to iterate quickly through different RL methods then it's a good idea to use one of the RL libraries like stable baselines 3. Then you can dig further into the methods that work best for you. Coding RL methods from scratch is very time consuming and error prone even for experienced programmers.
-
A note from our sponsor - InfluxDB
www.influxdata.com | 29 Mar 2024
Stats
DLR-RM/stable-baselines3 is an open source project licensed under MIT License which is an OSI approved license.
The primary programming language of stable-baselines3 is Python.
Popular Comparisons
- stable-baselines3 VS Ray
- stable-baselines3 VS stable-baselines
- stable-baselines3 VS Pytorch
- stable-baselines3 VS cleanrl
- stable-baselines3 VS tianshou
- stable-baselines3 VS Super-mario-bros-PPO-pytorch
- stable-baselines3 VS ElegantRL
- stable-baselines3 VS SuperSuit
- stable-baselines3 VS Tic-Tac-Toe-Gym
- stable-baselines3 VS rl-baselines3-zoo