yggdrasil-decision-forests
Spearmint
yggdrasil-decision-forests | Spearmint | |
---|---|---|
4 | 2 | |
428 | 1,529 | |
3.0% | 0.0% | |
9.5 | 0.0 | |
5 days ago | over 4 years ago | |
C++ | Python | |
Apache License 2.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
yggdrasil-decision-forests
-
Why do tree-based models still outperform deep learning on tabular data? (2022)
Is it this library https://github.com/google/yggdrasil-decision-forests ?
-
Binary image classification using random forest algorithm
However if you know cpp you can use Yggdrasil https://github.com/google/yggdrasil-decision-forests.
- Why do tree-based models still outperform deep learning on tabular data?
-
[P] Tree compiler that speeds up LightGBM model inference by ~30x
Have you tried to compare with Yggdrasil, the decision forest engine (c++, both training and inference) powering TensorFlow Decision Forests ?
Spearmint
-
Why do tree-based models still outperform deep learning on tabular data?
It occurs to me that a system, trained on peer-reviewed applied-machine-learning literature and Kaggle winners, that generates candidates for structured feature-engineering specifications, based on plaintext descriptions of columns' real-world meaning, should be considered a requisite part of the "meta" here.
Ah, and then you could iterate within the resulting feature-engineering-suggestion space as a hyper-parameter between experiments, which could be optimized with e.g. https://github.com/HIPS/Spearmint . The papers write themselves!
-
[D] What kind of Hyperparameter Optimisation do you use?
This was some time ago but I had some promising results with Bayesian optimization using a Gaussian Process prior. The method was developed by the guys who wrote Spearmint. That library doesn't support parallelization but I implemented the same technique in Scala without too much difficulty.
What are some alternatives?
LightGBM - A fast, distributed, high performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.
optuna - A hyperparameter optimization framework
tensorflow - An Open Source Machine Learning Framework for Everyone
srbench - A living benchmark framework for symbolic regression
decision-tree-classifier - Decision Tree Classifier and Boosted Random Forest
axe-testcafe - The helper for using Axe in TestCafe tests
flashlight - A C++ standalone library for machine learning [Moved to: https://github.com/flashlight/flashlight]
youtube-react - A Youtube clone built in React, Redux, Redux-saga
interpret - Fit interpretable models. Explain blackbox machine learning.
spaceopt - Hyperparameter optimization via gradient boosting regression
decision-forests - A collection of state-of-the-art algorithms for the training, serving and interpretation of Decision Forest models in Keras.