SaaSHub helps you find the best software and product alternatives Learn more →
Stable-baselines3 Alternatives
Similar projects and alternatives to stable-baselines3
-
Ray
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
-
stable-baselines
A fork of OpenAI Baselines, implementations of reinforcement learning algorithms
-
Onboard AI
Learn any GitHub repo in 59 seconds. Onboard AI learns any GitHub repo in minutes and lets you chat with it to locate functionality, understand different parts, and generate new code. Use it for free at www.getonboard.dev.
-
cleanrl
High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)
-
-
Pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
-
PettingZoo
An API standard for multi-agent reinforcement learning environments, with popular reference environments and related utilities
-
-
InfluxDB
Collect and Analyze Billions of Data Points in Real Time. Manage all types of time series data in a single, purpose-built database. Run at any scale in any environment in the cloud, on-premises, or at the edge.
-
Super-mario-bros-PPO-pytorch
Proximal Policy Optimization (PPO) algorithm for Super Mario Bros
-
rl-baselines3-zoo
A training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included.
-
SuperSuit
A collection of wrappers for Gymnasium and PettingZoo environments (being merged into gymnasium.wrappers and pettingzoo.wrappers
-
stable-baselines3-contrib
Contrib package for Stable-Baselines3 - Experimental reinforcement learning (RL) code
-
gym
A toolkit for developing and comparing reinforcement learning algorithms.
-
baselines
OpenAI Baselines: high-quality implementations of reinforcement learning algorithms
-
Deep-Reinforcement-Learning-Algorithms-with-PyTorch
PyTorch implementations of deep reinforcement learning algorithms and environments
-
pytorch-trpo
PyTorch implementation of Trust Region Policy Optimization
-
machin
Reinforcement learning library(framework) designed for PyTorch, implements DQN, DDPG, A2C, PPO, SAC, MADDPG, A3C, APEX, IMPALA ...
-
Tic-Tac-Toe-Gym
This is the Tic-Tac-Toe game made with Python using the PyGame library and the Gym library to implement the AI with Reinforcement Learning
-
-
RL-Adventure
Pytorch Implementation of DQN / DDQN / Prioritized replay/ noisy networks/ distributional values/ Rainbow/ hierarchical RL
-
agents
TF-Agents: A reliable, scalable and easy to use TensorFlow library for Contextual Bandits and Reinforcement Learning.
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
stable-baselines3 reviews and mentions
-
[P] PettingZoo 1.24.0 has been released (including Stable-Baselines3 tutorials)
PettingZoo 1.24.0 is now live! This release includes Python 3.11 support, updated Chess and Hanabi environment versions, and many bugfixes, documentation updates and testing expansions. We are also very excited to announce 3 tutorials using Stable-Baselines3, and a full training script using CleanRL with TensorBoard and WandB.
-
SB3 - NotImplementedError: Box([-1. -1. -8.], [1. 1. 8.], (3,), <class 'numpy.float32'>) observation space is not supported
Therefore, I debugged this error to the ReplayBuffer that was imported from `SB3`. This is the problem function -
-
Shimmy 1.0: Gymnasium & PettingZoo bindings for popular external RL environments
Have you ever wanted to use dm-control with stable-baselines3? Within Reinforcement learning (RL), a number of APIs are used to implement environments, with limited ability to convert between them. This makes training agents across different APIs highly difficult, and has resulted in a fractured ecosystem.
-
Stable-Baselines3 v1.8 Release
Changelog: https://github.com/DLR-RM/stable-baselines3/releases/tag/v1.8.0
-
[P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up
Great project! One question though, is there any reason why you are not using existing RL models instead of creating your own, such as stable baselines?
-
Is Stable Baselines 3 no longer compatible with PettingZoo?
I was able to get Stable Baselines 3 to work with gymnasium by following the details in this work-in-progress PR: https://github.com/DLR-RM/stable-baselines3/pull/780. I have not used PettingZoo, though.
- [OC] Inteligência Artificial jogando Mega Man!
-
New to reinforcement learning.
I'd say this is a great path but I'd also look at the basic on-policy gradient actor critic methods like A2C and eventually PPO. Someone recommended SAC which also really good. There are tons of environments in the https://github.com/Farama-Foundation/PettingZoo as well if you want to mess with those. You can also check out stable baselines https://github.com/DLR-RM/stable-baselines3 which is pretty popular. If you want to get into the theory more I recommend reading the Sutton and Barto book on reinforcement learning.
-
How to proceed further? (Learning RL)
If you want to iterate quickly through different RL methods then it's a good idea to use one of the RL libraries like stable baselines 3. Then you can dig further into the methods that work best for you. Coding RL methods from scratch is very time consuming and error prone even for experienced programmers.
-
PPO rollout buffer for turn-based two-player game with varying turn lengths
Algorithm to compute returns and advantages (I use MaskableDictRolloutBuffer which inherits the shown implementation from RolloutBuffer (line 371)):
-
A note from our sponsor - #<SponsorshipServiceOld:0x00007f0f9bb28228>
www.saashub.com | 5 Dec 2023
Stats
DLR-RM/stable-baselines3 is an open source project licensed under MIT License which is an OSI approved license.
The primary programming language of stable-baselines3 is Python.
Popular Comparisons
- stable-baselines3 VS Ray
- stable-baselines3 VS stable-baselines
- stable-baselines3 VS Pytorch
- stable-baselines3 VS tianshou
- stable-baselines3 VS cleanrl
- stable-baselines3 VS Super-mario-bros-PPO-pytorch
- stable-baselines3 VS ElegantRL
- stable-baselines3 VS SuperSuit
- stable-baselines3 VS Tic-Tac-Toe-Gym
- stable-baselines3 VS rl-baselines3-zoo