cuml
scikit-learn
Our great sponsors
cuml | scikit-learn | |
---|---|---|
10 | 81 | |
3,894 | 58,046 | |
2.0% | 1.0% | |
9.3 | 9.9 | |
about 17 hours ago | 3 days ago | |
C++ | Python | |
Apache License 2.0 | BSD 3-clause "New" or "Revised" License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
cuml
- FLaNK Stack Weekly for 13 November 2023
-
Is it possible to run Sklearn models on a GPU?
sklearn can't, bit take a look at cuML (https://github.com/rapidsai/cuml ). It uses the same API as sklearn but executes on GPU.
-
[P] Looking for state of the art clustering algorithms
As a companion to the other comments, I'd like to mention that the RAPIDS library cuML provides GPU-accelerated versions of quite a few of the algorithms mentioned in this thread (HDBSCAN, UMAP, SVM, PCA, {Exact, Approximate} Nearest Neighbors, DBSCAN, KMeans, etc.).
-
Is there a multi regression model that works on GPU?
CuML
- [D] What's your favorite unpopular/forgotten Machine Learning method?
- Machine Learning with PyTorch and Scikit-Learn – The *New* Python ML Book
-
What are the advantages and disadvantages of using GPU for machine learning/ deep learning/ scientific computation over the conventional CPU software acceleration?
Did they implement the clustering algorithm themselves? cuML is a GPU-accelerated scikit-learn-like package that covers many of the common ML algorithms.
-
Intel Extension for Scikit-Learn
https://github.com/rapidsai/cuml
> cuML is a suite of libraries that implement machine learning algorithms and mathematical primitives functions that share compatible APIs with other RAPIDS projects. cuML enables data scientists, researchers, and software engineers to run traditional tabular ML tasks on GPUs without going into the details of CUDA programming. In most cases, cuML's Python API matches the API from scikit-learn. For large datasets, these GPU-based implementations can complete 10-50x faster than their CPU equivalents. For details on performance, see the cuML Benchmarks Notebook.
-
GPU Based Kernel-PCA
Cython code
-
Python Machine Learning Guy getting started with CUDA. What should I be brushing up on?
Take a look at RAPIDS CUML https://github.com/rapidsai/cuml. It's useful for most common ML algorithms. Feel free to create Github issues for feature requests & bugs.
scikit-learn
-
AutoCodeRover resolves 22% of real-world GitHub in SWE-bench lite
Thank you for your interest. There are some interesting examples in the SWE-bench-lite benchmark which are resolved by AutoCodeRover:
- From sympy: https://github.com/sympy/sympy/issues/13643. AutoCodeRover's patch for it: https://github.com/nus-apr/auto-code-rover/blob/main/results...
- Another one from scikit-learn: https://github.com/scikit-learn/scikit-learn/issues/13070. AutoCodeRover's patch (https://github.com/nus-apr/auto-code-rover/blob/main/results...) modified a few lines below (compared to the developer patch) and wrote a different comment.
There are more examples in the results directory (https://github.com/nus-apr/auto-code-rover/tree/main/results).
-
Polars
sklearn is adding support through the dataframe interchange protocol (https://github.com/scikit-learn/scikit-learn/issues/25896). scipy, as far as I know, doesn't explicitly support dataframes (it just happens to work when you wrap a Series in `np.array` or `np.asarray`). I don't know about PyTorch but in general you can convert to numpy.
-
[D] Major bug in Scikit-Learn's implementation of F-1 score
Wow, from the upvotes on this comment, it really seems like a lot of people think that this is the correct behavior! I have to say I disagree, but if that's what you think, don't just sit there upvoting comments on Reddit; instead go to this PR and tell the Scikit-Learn maintainers not to "fix" this "bug", which they are currently planning to do!
- Contraction Clustering (RASTER): A fast clustering algorithm
-
Ask HN: Learning new coding patterns – how to start?
I was in a similar boat to yours - Worked in data science and since then have made a move to data engineering and software engineering for ML services.
I would recommend you look into the Design Patterns book by the Gang of Four. I found it particularly helpful to make extensible code that doesn't break specially with abstract classes, builders and factories. I would also recommend looking into the book The Object Oriented Thought Process to understand why traditional OOP is build the way it is.
You can also look into the source code of popular data science libraries such as sklearn (https://github.com/scikit-learn/scikit-learn/tree/main/sklea...) and see how a lot of them have Base classes to define shared functionality between object of the same nature.
As others mentioned, I would also encourage you to try and implement design patterns in your everyday work - maybe you can make a Factory to load models or preprocessors that follow the same Abstract class?
-
Transformers as Support Vector Machines
It looks like you've been the victim of some misinformation. As Dr_Birdbrain said, an SVM is a convex problem with unique global optimum. sklearn.SVC relies on libsvm which initializes the weights to 0 [0]. The random state is only used to shuffle the data to make probability estimates with Platt scaling [1]. Of the random_state parameter, the sklearn documentation for SVC [2] says
Controls the pseudo random number generation for shuffling the data for probability estimates. Ignored when probability is False. Pass an int for reproducible output across multiple function calls. See Glossary.
[0] https://github.com/scikit-learn/scikit-learn/blob/2a2772a87b...
[1] https://en.wikipedia.org/wiki/Platt_scaling
[2] https://scikit-learn.org/stable/modules/generated/sklearn.sv...
-
How to Build and Deploy a Machine Learning model using Docker
Scikit-learn Documentation
- Planning to get a laptop for ML/DL, is this good enough at the price point or are there better options at/below this price point?
-
Link Prediction With node2vec in Physics Collaboration Network
Firstly, we need a connection to Memgraph so we can get edges, split them into two parts (train set and test set). For edge splitting, we will use scikit-learn. In order to make a connection towards Memgraph, we will use gqlalchemy.
-
WiFilter is a RaspAP install extended with a squidGuard proxy to filter adult content. Great solution for a family, schools and/or public access point
The ML component is based on scikit-learn which differentiates it from purely list-based filters. It couples this with a full-featured wireless router (RaspAP) in a single device, so it fulfills the needs of a use case not entirely addressed by Pi-hole.
What are some alternatives?
scikit-learn-intelex - Intel(R) Extension for Scikit-learn is a seamless way to speed up your Scikit-learn application
Prophet - Tool for producing high quality forecasts for time series data that has multiple seasonality with linear or non-linear growth.
scikit-cuda - Python interface to GPU-powered libraries
Surprise - A Python scikit for building and analyzing recommender systems
hummingbird - Hummingbird compiles trained ML models into tensor computation for faster inference.
Keras - Deep Learning for humans
cudf - cuDF - GPU DataFrame Library
tensorflow - An Open Source Machine Learning Framework for Everyone
evojax
gensim - Topic Modelling for Humans
lightseq - LightSeq: A High Performance Library for Sequence Processing and Generation
H2O - H2O is an Open Source, Distributed, Fast & Scalable Machine Learning Platform: Deep Learning, Gradient Boosting (GBM) & XGBoost, Random Forest, Generalized Linear Modeling (GLM with Elastic Net), K-Means, PCA, Generalized Additive Models (GAM), RuleFit, Support Vector Machine (SVM), Stacked Ensembles, Automatic Machine Learning (AutoML), etc.