Intel Extension for Scikit-Learn

This page summarizes the projects mentioned and recommended in the original post on news.ycombinator.com

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • scikit-learn

    scikit-learn: machine learning in Python

  • Hi all,

    Currently some works is being done to improve computational primitives of scikit-learn to enhance its overhaul performances natively.

    You can have a look at this exploratory PR: https://github.com/scikit-learn/scikit-learn/pull/20254

    This other PR is a clear revamp of this previous one:

  • scikit-learn-intelex

    Intel(R) Extension for Scikit-learn is a seamless way to speed up your Scikit-learn application

  • Looks like they are responding to https://github.com/intel/scikit-learn-intelex#-acceleration

    I completely agree. I hope some Intel competitor funds a scikit-learn developer to read this code and extract all the portable performance improvements.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • cuml

    cuML - RAPIDS Machine Learning Library

  • https://github.com/rapidsai/cuml

    > cuML is a suite of libraries that implement machine learning algorithms and mathematical primitives functions that share compatible APIs with other RAPIDS projects. cuML enables data scientists, researchers, and software engineers to run traditional tabular ML tasks on GPUs without going into the details of CUDA programming. In most cases, cuML's Python API matches the API from scikit-learn. For large datasets, these GPU-based implementations can complete 10-50x faster than their CPU equivalents. For details on performance, see the cuML Benchmarks Notebook.

  • numba-dpex

    Data Parallel Extension for Numba

  • > Intel are focused on data-parallel C++ for delivering high performance, rightly or wrongly.

    They also invest efforts in making it possible to write high performance kernels in Python using an extension to the numba Python compiler:

    https://github.com/IntelPython/numba-dppy

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • How to Build a Logistic Regression Model: A Spam-filter Tutorial

    1 project | dev.to | 5 May 2024
  • [D] Major bug in Scikit-Learn's implementation of F-1 score

    2 projects | /r/MachineLearning | 8 Dec 2023
  • Contraction Clustering (RASTER): A fast clustering algorithm

    1 project | news.ycombinator.com | 27 Nov 2023
  • Transformers as Support Vector Machines

    1 project | news.ycombinator.com | 3 Sep 2023
  • How to Build and Deploy a Machine Learning model using Docker

    5 projects | dev.to | 30 Jul 2023