Our great sponsors
-
xgboost
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
By the way, most of the time XGBoost works just as well for projects, would not recommend applying deep learning to every single problem you come across, it's something Stanford CS really likes to showcase when it's well known (1) that sometimes "smaller"/less complex models can perform just as well or have their own interpretive advantages and (2) it is well known within ML and DS communities that deep learning does not perform as well with tabular datasets and using deep learning as a default to every problem is just poor practice. However, if you do (god forbid) get language, speech/audio, vision/imaging, or even time series models then deep learning as a baseline is not the worst idea.
Related posts
- OOM with ML Models (SKlearn, XGBoost, etc), workaround/tips for large datasets?
- 'y contains previously unseen labels' (label encoder)
- What's New with AWS: Amazon SageMaker built-in algorithms now provides four new Tabular Data Modeling Algorithms
-
xgboost VS CXXGraph - a user suggested alternative
2 projects | 28 Feb 2022
-
LightGBM VS CXXGraph - a user suggested alternative
2 projects | 28 Feb 2022