neptune-contrib
Sacred
Our great sponsors
neptune-contrib | Sacred | |
---|---|---|
- | 6 | |
27 | 4,155 | |
- | 0.4% | |
0.0 | 3.5 | |
over 1 year ago | 2 months ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
neptune-contrib
We haven't tracked posts mentioning neptune-contrib yet.
Tracking mentions began in Dec 2020.
Sacred
-
Sacred VS cascade - a user suggested alternative
2 projects | 5 Dec 2023
-
✨ 7 Best Machine Learning Experiment Logging Tools in 2022 🚀
🔗 https://github.com/IDSIA/sacred
-
https://np.reddit.com/r/MachineLearning/comments/pvs8r5/d_facebook_visdom_vs_google_tensorboard_for/hefg131/
I'm using Omniboard (https://github.com/vivekratnavel/omniboard) with Sacred (https://github.com/IDSIA/sacred) for tracking experiments. You can specify custom Observers in Sacred so the model metrics and logs will be saved to a local directory or to a remote DB (e.g., MongoDB). I use a MongoDB database hosted on Atlas. Unlike other suggested options, Sacred and Omniboard are free. Atlas free tier comes with 512MB of free storage which is a huge amount if you're uploading only log files to it.
-
[D] Facebook Visdom vs Google Tensorboard for Pytorch
I'm using Omniboard (https://github.com/vivekratnavel/omniboard) with Sacred (https://github.com/IDSIA/sacred) for tracking experiments. You can specify custom Observers in Sacred so the model metrics and logs will be saved to a local directory or to a remote DB (e.g., MongoDB). I use a MongoDB database hosted on Atlas. Unlike other suggested options, Sacred and Omniboard are free. Atlas free tier comes with 512MB of free storage which is a huge amount if you're uploading only log files to it. ex = Experiment() ex.observers.append(FileStorageObserver(EXPERIMENTS_ROOT)) ex.observers.append(MongoObserver(url=MONGODB_URL, db_name='sacred'))
-
Can someone tell me good libraries you use on a day to day basis that increases your research productivity in ML/AI?
sacred helped me log my experiments. I did setup my environment only once 4 years ago, and since then I have a list of all my training runs with the hyperparameters and results.
-
[D] How to be more productive while doing Deep Learning experiments?
For 1, setup an experiment tracking framework. I found Sacred to be helpful https://github.com/IDSIA/sacred.
What are some alternatives?
scikit-learn - scikit-learn: machine learning in Python
MLflow - Open source platform for the machine learning lifecycle
tensorflow - An Open Source Machine Learning Framework for Everyone
pytorch-lightning - Build high-performance AI models with PyTorch Lightning (organized PyTorch). Deploy models with Lightning Apps (organized Python to build end-to-end ML systems). [Moved to: https://github.com/Lightning-AI/lightning]
Keras - Deep Learning for humans
bodywork - ML pipeline orchestration and model deployments on Kubernetes.
xgboost - Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
Clairvoyant - Software designed to identify and monitor social/historical cues for short term stock movement
TFLearn - Deep learning library featuring a higher-level API for TensorFlow.