Sacred
gym
Our great sponsors
Sacred | gym | |
---|---|---|
6 | 96 | |
4,157 | 33,873 | |
0.4% | 0.8% | |
3.5 | 0.0 | |
2 months ago | 20 days ago | |
Python | Python | |
MIT License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Sacred
-
Sacred VS cascade - a user suggested alternative
2 projects | 5 Dec 2023
-
✨ 7 Best Machine Learning Experiment Logging Tools in 2022 🚀
🔗 https://github.com/IDSIA/sacred
-
https://np.reddit.com/r/MachineLearning/comments/pvs8r5/d_facebook_visdom_vs_google_tensorboard_for/hefg131/
I'm using Omniboard (https://github.com/vivekratnavel/omniboard) with Sacred (https://github.com/IDSIA/sacred) for tracking experiments. You can specify custom Observers in Sacred so the model metrics and logs will be saved to a local directory or to a remote DB (e.g., MongoDB). I use a MongoDB database hosted on Atlas. Unlike other suggested options, Sacred and Omniboard are free. Atlas free tier comes with 512MB of free storage which is a huge amount if you're uploading only log files to it.
-
[D] Facebook Visdom vs Google Tensorboard for Pytorch
I'm using Omniboard (https://github.com/vivekratnavel/omniboard) with Sacred (https://github.com/IDSIA/sacred) for tracking experiments. You can specify custom Observers in Sacred so the model metrics and logs will be saved to a local directory or to a remote DB (e.g., MongoDB). I use a MongoDB database hosted on Atlas. Unlike other suggested options, Sacred and Omniboard are free. Atlas free tier comes with 512MB of free storage which is a huge amount if you're uploading only log files to it. ex = Experiment() ex.observers.append(FileStorageObserver(EXPERIMENTS_ROOT)) ex.observers.append(MongoObserver(url=MONGODB_URL, db_name='sacred'))
-
Can someone tell me good libraries you use on a day to day basis that increases your research productivity in ML/AI?
sacred helped me log my experiments. I did setup my environment only once 4 years ago, and since then I have a list of all my training runs with the hyperparameters and results.
-
[D] How to be more productive while doing Deep Learning experiments?
For 1, setup an experiment tracking framework. I found Sacred to be helpful https://github.com/IDSIA/sacred.
gym
-
OpenAI Acquires Global Illumination
A co-founder announced they disbanded their robots team a couple years ago: https://venturebeat.com/business/openai-disbands-its-robotic...
That was the same time they depreciated OpenAI Gym: https://github.com/openai/gym
-
Shimmy 1.0: Gymnasium & PettingZoo bindings for popular external RL environments
This includes single-agent Gymnasium wrappers for DM Control, DM Lab, Behavior Suite, Arcade Learning Environment, OpenAI Gym V21 & V26. Multi-agent PettingZoo wrappers support DM Control Soccer, OpenSpiel and Melting Pot. For more information, read the release notes here:
-
Some confusion about variables and functions in mujoco-py
When I browse fetch_env.py, I have a question about the following code snippet:
-
pip install stable-baselines3[extra]
Nvm, this works for me '!pip install setuptools==65.5.0' Source: https://github.com/openai/gym/issues/3176
-
[P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up
how would this interact/compare with https://github.com/openai/gym?
- What has replaced OpenAI Retro Gym?
-
Understanding Reinforcement Learning
If you'd like to learn more about reinforcement learning or play with a number of samples in controlled environments, I highly recommend you look at the documentation for OpenAI's Gym library and particularly the basic usage page. OpenAI's Gym provides a standardized environment for performing reinforcement learning on classic Atari games and a few other platforms and should be an educational resource. If you'd like a more detailed example, check out this tutorial on Paperspace's blog.
-
Using the cross-entropy method to solve Frozen Lake
Frozen Lake is an OpenAI Gym environment in which an agent is rewarded for traversing a frozen surface from a start position to a goal position without falling through any perilous holes in the ice.
- Is there a publicly available state space model for the Lunar Lander environment?
-
How to Create a Behavioral Cloning Bot to Play Online Games?
typically a more relaxed approach is taken via reinforcement learning, but it requires that you can simulate the game via a given gamestate. take a look at e.g. https://www.gymlibrary.dev/
What are some alternatives?
MLflow - Open source platform for the machine learning lifecycle
ml-agents - The Unity Machine Learning Agents Toolkit (ML-Agents) is an open-source project that enables games and simulations to serve as environments for training intelligent agents using deep reinforcement learning and imitation learning.
pytorch-lightning - Build high-performance AI models with PyTorch Lightning (organized PyTorch). Deploy models with Lightning Apps (organized Python to build end-to-end ML systems). [Moved to: https://github.com/Lightning-AI/lightning]
carla - Open-source simulator for autonomous driving research.
tensorflow - An Open Source Machine Learning Framework for Everyone
Keras - Deep Learning for humans
dm_control - Google DeepMind's software stack for physics-based simulation and Reinforcement Learning environments, using MuJoCo.
Clairvoyant - Software designed to identify and monitor social/historical cues for short term stock movement
open_spiel - OpenSpiel is a collection of environments and algorithms for research in general reinforcement learning and search/planning in games.
scikit-learn - scikit-learn: machine learning in Python
rlcard - Reinforcement Learning / AI Bots in Card (Poker) Games - Blackjack, Leduc, Texas, DouDizhu, Mahjong, UNO.