SHOGUN VS xgboost

Compare SHOGUN vs xgboost and see what are their differences.

xgboost

Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow (by dmlc)
Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
SHOGUN xgboost
1 10
3,005 25,548
0.5% 0.9%
4.8 9.6
4 months ago 6 days ago
C++ C++
BSD 3-clause "New" or "Revised" License Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

SHOGUN

Posts with mentions or reviews of SHOGUN. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-04-20.
  • Changing std:sort at Google’s Scale and Beyond
    7 projects | news.ycombinator.com | 20 Apr 2022
    The function is trying to get the median, which is not defined for an empty set. With this particular implementation, there is an assert for that:

    https://github.com/shogun-toolbox/shogun/blob/9b8d85/src/sho...

    Unrelatedly, but from the same section:

    > Fixes are trivial, access the nth element only after the call being made. Be careful.

    Wouldn't the proper fix to do the nth_element for the larget element first (for those cases that don't do that already) and then adjust the end to be the begin + larger_n for the second nth_element call? Otherwise the second call will check [begin + larger_n, end) again for no reason at all.

xgboost

Posts with mentions or reviews of xgboost. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-05-09.

What are some alternatives?

When comparing SHOGUN and xgboost you can also consider the following projects:

mlpack - mlpack: a fast, header-only C++ machine learning library

Prophet - Tool for producing high quality forecasts for time series data that has multiple seasonality with linear or non-linear growth.

Caffe - Caffe: a fast open framework for deep learning.

MLP Classifier - A handwritten multilayer perceptron classifer using numpy.

mxnet - Lightweight, Portable, Flexible Distributed/Mobile Deep Learning with Dynamic, Mutation-aware Dataflow Dep Scheduler; for Python, R, Julia, Scala, Go, Javascript and more

tensorflow - An Open Source Machine Learning Framework for Everyone

Dlib - A toolkit for making real world machine learning and data analysis applications in C++

Keras - Deep Learning for humans

vowpal_wabbit - Vowpal Wabbit is a machine learning system which pushes the frontier of machine learning with techniques such as online, hashing, allreduce, reductions, learning2search, active, and interactive learning.

OpenHotspot

catboost - A fast, scalable, high performance Gradient Boosting on Decision Trees library, used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. Supports computation on CPU and GPU.