Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Top 8 Python stable-baseline Projects
-
stable-baselines3
PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.
-
rl-baselines3-zoo
A training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
rl-baselines-zoo
A collection of 100+ pre-trained RL agents using Stable Baselines, training and hyperparameter optimization included.
-
stable-baselines3-contrib
Contrib package for Stable-Baselines3 - Experimental reinforcement learning (RL) code
-
learning-to-drive-in-5-minutes
Implementation of reinforcement learning approach to make a car learn to drive smoothly in minutes
-
stable-baselines
Mirror of Stable-Baselines: a fork of OpenAI Baselines, implementations of reinforcement learning algorithms (by Stable-Baselines-Team)
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
Tic-Tac-Toe-Gym
This is the Tic-Tac-Toe game made with Python using the PyGame library and the Gym library to implement the AI with Reinforcement Learning
The latest release (v3.0.0) of Upkie's software brings a functional sim-to-real reinforcement learning pipeline based on Stable Baselines3, with standard sim-to-real tricks. The pipeline trains on the Gymnasium environments distributed in upkie.envs (setup: pip install upkie) and is implemented in the PPO balancer. Here is a policy running on an Upkie:
Project mention: Can't solve MountainCar-v0 with A2C algorithm (stable-baselines3) | /r/reinforcementlearning | 2023-06-27I'm trying to solve MountainCar-v0 enviroment from gymnasium with the A2C algorithm and the agent doesn't find a solution. I checked this so I added import stable_baselines3.common.sb2_compat.rmsprop_tf_like as RMSpropTFLike. Also checked the rl-baselines3-zoo for the hyperparameter tuning. So my code is:
Project mention: Problem with Truncated Quantile Critics (TQC) and n-step learning algorithm. | /r/reinforcementlearning | 2023-12-09# https://github.com/Stable-Baselines-Team/stable-baselines3-contrib/blob/master/sb3_contrib/tqc/tqc.py :
Python stable-baselines related posts
-
Problem with Truncated Quantile Critics (TQC) and n-step learning algorithm.
-
Can't solve MountainCar-v0 with A2C algorithm (stable-baselines3)
-
Stable-Baselines3 v2.0: Gymnasium Support
-
Understanding Action Masking in RLlib
-
Agent trains great with PPO but terrible with SAC --> Advice for Hyperparameters
-
PPO rollout buffer for turn-based two-player game with varying turn lengths
-
Where can I get pre trained machine learning models?
-
A note from our sponsor - InfluxDB
www.influxdata.com | 10 May 2024
Index
What are some of the best open-source stable-baseline projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | stable-baselines3 | 7,988 |
2 | rl-baselines3-zoo | 1,791 |
3 | rl-baselines-zoo | 1,106 |
4 | stable-baselines3-contrib | 429 |
5 | learning-to-drive-in-5-minutes | 277 |
6 | stable-baselines | 277 |
7 | rl-trained-agents | 93 |
8 | Tic-Tac-Toe-Gym | 8 |
Sponsored