Sacred
xgboost
Our great sponsors
Sacred | xgboost | |
---|---|---|
6 | 10 | |
4,157 | 25,576 | |
0.4% | 0.9% | |
3.5 | 9.6 | |
2 months ago | about 9 hours ago | |
Python | C++ | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Sacred
-
Sacred VS cascade - a user suggested alternative
2 projects | 5 Dec 2023
-
✨ 7 Best Machine Learning Experiment Logging Tools in 2022 🚀
🔗 https://github.com/IDSIA/sacred
-
https://np.reddit.com/r/MachineLearning/comments/pvs8r5/d_facebook_visdom_vs_google_tensorboard_for/hefg131/
I'm using Omniboard (https://github.com/vivekratnavel/omniboard) with Sacred (https://github.com/IDSIA/sacred) for tracking experiments. You can specify custom Observers in Sacred so the model metrics and logs will be saved to a local directory or to a remote DB (e.g., MongoDB). I use a MongoDB database hosted on Atlas. Unlike other suggested options, Sacred and Omniboard are free. Atlas free tier comes with 512MB of free storage which is a huge amount if you're uploading only log files to it.
-
[D] Facebook Visdom vs Google Tensorboard for Pytorch
I'm using Omniboard (https://github.com/vivekratnavel/omniboard) with Sacred (https://github.com/IDSIA/sacred) for tracking experiments. You can specify custom Observers in Sacred so the model metrics and logs will be saved to a local directory or to a remote DB (e.g., MongoDB). I use a MongoDB database hosted on Atlas. Unlike other suggested options, Sacred and Omniboard are free. Atlas free tier comes with 512MB of free storage which is a huge amount if you're uploading only log files to it. ex = Experiment() ex.observers.append(FileStorageObserver(EXPERIMENTS_ROOT)) ex.observers.append(MongoObserver(url=MONGODB_URL, db_name='sacred'))
-
Can someone tell me good libraries you use on a day to day basis that increases your research productivity in ML/AI?
sacred helped me log my experiments. I did setup my environment only once 4 years ago, and since then I have a list of all my training runs with the hyperparameters and results.
-
[D] How to be more productive while doing Deep Learning experiments?
For 1, setup an experiment tracking framework. I found Sacred to be helpful https://github.com/IDSIA/sacred.
xgboost
- XGBoost 2.0
- XGBoost2.0
- Xgboost: Banding continuous variables vs keeping raw data
-
PSA: You don't need fancy stuff to do good work.
Finally, when it comes to building models and making predictions, Python and R have a plethora of options available. Libraries like scikit-learn, statsmodels, and TensorFlowin Python, or caret, randomForest, and xgboostin R, provide powerful machine learning algorithms and statistical models that can be applied to a wide range of problems. What's more, these libraries are open-source and have extensive documentation and community support, making it easy to learn and apply new techniques without needing specialized training or expensive software licenses.
-
XGBoost Save and Load Error
You can find the problem outlined here: https://github.com/dmlc/xgboost/issues/5826. u/hcho3 diagnosed the problem and corrected it as of XGB version 1.2.0.
-
For XGBoost (in Amazon SageMaker), one of the hyper parameters is num_round, for number of rounds to train. Does this mean cross validation?
Reference: https://github.com/dmlc/xgboost/issues/2031
-
CS Internship Questions
By the way, most of the time XGBoost works just as well for projects, would not recommend applying deep learning to every single problem you come across, it's something Stanford CS really likes to showcase when it's well known (1) that sometimes "smaller"/less complex models can perform just as well or have their own interpretive advantages and (2) it is well known within ML and DS communities that deep learning does not perform as well with tabular datasets and using deep learning as a default to every problem is just poor practice. However, if you do (god forbid) get language, speech/audio, vision/imaging, or even time series models then deep learning as a baseline is not the worst idea.
- OOM with ML Models (SKlearn, XGBoost, etc), workaround/tips for large datasets?
-
xgboost VS CXXGraph - a user suggested alternative
2 projects | 28 Feb 2022
- 'y contains previously unseen labels' (label encoder)
What are some alternatives?
MLflow - Open source platform for the machine learning lifecycle
Prophet - Tool for producing high quality forecasts for time series data that has multiple seasonality with linear or non-linear growth.
pytorch-lightning - Build high-performance AI models with PyTorch Lightning (organized PyTorch). Deploy models with Lightning Apps (organized Python to build end-to-end ML systems). [Moved to: https://github.com/Lightning-AI/lightning]
MLP Classifier - A handwritten multilayer perceptron classifer using numpy.
tensorflow - An Open Source Machine Learning Framework for Everyone
Keras - Deep Learning for humans
Clairvoyant - Software designed to identify and monitor social/historical cues for short term stock movement
mlpack - mlpack: a fast, header-only C++ machine learning library
scikit-learn - scikit-learn: machine learning in Python
catboost - A fast, scalable, high performance Gradient Boosting on Decision Trees library, used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. Supports computation on CPU and GPU.