Youtube-Code-Repository
Repository for most of the code from my YouTube channel (by philtabor)
minimalRL
Implementations of basic RL algorithms with minimal lines of codes! (pytorch based) (by seungeunrho)
Youtube-Code-Repository | minimalRL | |
---|---|---|
5 | 5 | |
844 | 2,725 | |
- | - | |
1.6 | 1.6 | |
10 months ago | about 1 year ago | |
Python | Python | |
- | MIT License |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Youtube-Code-Repository
Posts with mentions or reviews of Youtube-Code-Repository.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-04-28.
-
Overall loss in PPO, why does it matter?
In Phil tabor's implementation it calculates Actor and Critic loss separately (line 95+) and does not calculate equation 9.
-
Intrinsic Curiosity Module Pytorch multithreading cpu unable to fix seeds
I am working on an extension of this implementation https://github.com/philtabor/Youtube-Code-Repository/tree/master/ReinforcementLearning/ICM of the intrinsic curiosity module. It uses A3C(Actor -critic) as a policy and the ICM is a bolt on module.
-
PPO cannot play CartPole ?
A very good performance reference code, which convers in 200 episodes.
-
Rl algorithm implemented
Github code - https://github.com/philtabor/Youtube-Code-Repository/tree/master/ReinforcementLearning/PolicyGradient/DDPG/tensorflow2/pendulum
-
Lunar Lander using Deep Q-Learning
I was wondering why the code looked so familiar, not just the design, but even the syntax and names of functions. I went through these myself when I was learning: Youtube-Code-Repository/ReinforcementLearning/DeepQLearning at master · philtabor/Youtube-Code-Repository (github.com). Its by a YouTuber / Udemy course instructor that goes through the design and coding process from scratch. This is probably mostly lifted straight from that repo. He even has a video on doing the lunar lander example too.
minimalRL
Posts with mentions or reviews of minimalRL.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2021-07-18.
- Does anyone know good python sources hardcoded of RL?
-
Question about pseudocodes
Did you try minimalRL?
- Rl algorithm implemented
-
RL agent for simple games?
This github is great.
-
PPO+LSTM Implementation
Maybe this implementation helps: https://github.com/seungeunrho/minimalRL/blob/master/ppo-lstm.py
What are some alternatives?
When comparing Youtube-Code-Repository and minimalRL you can also consider the following projects:
Respiratory-Disease-Coughing-Dataset-CNN - A collection of coughing audio files from Coswara, Coughvid, and Virufy as well as generated spectrograms for the use of machine learning
ElegantRL - Massively Parallel Deep Reinforcement Learning. 🔥