Onboard AI learns any GitHub repo in minutes and lets you chat with it to locate functionality, understand different parts, and generate new code. Use it for free at www.getonboard.dev. Learn more →
Top 23 C++ Cuda Projects
Open3D: A Modern Library for 3D Data ProcessingProject mention: Does anyone else agree that the links to the latest development version of Open3D don't work? | /r/cscareerquestions | 2023-07-10
I was going to file a bug about another issue, but I have to download the development version. This is why I want this solved quickly. None of the links seem to work: https://github.com/isl-org/Open3D/issues/6259
cuDF - GPU DataFrame LibraryProject mention: A Polars exploration into Kedro | dev.to | 2023-05-17
The interesting thing about Polars is that it does not try to be a drop-in replacement to pandas, like Dask, cuDF, or Modin, and instead has its own expressive API. Despite being a young project, it quickly got popular thanks to its easy installation process and its “lightning fast” performance.
Collect and Analyze Billions of Data Points in Real Time. Manage all types of time series data in a single, purpose-built database. Run at any scale in any environment in the cloud, on-premises, or at the edge.
OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient.Project mention: OneFlow v0.9.0 Came Out!——A Distributed Deep Learning Framework | /r/programming | 2023-02-12
The C++ parallel algorithms library.
ArrayFire: a general purpose GPU library.Project mention: Learn WebGPU | news.ycombinator.com | 2023-04-27
Loads of people have stated why easy GPU interfaces are difficult to create, but we solve many difficult things all the time.
Ultimately I think CPUs are just satisfactory for the vast vast majority of workloads. Servers rarely come with any GPUs to speak of. The ecosystem around GPUs is unattractive. CPUs have SIMD instructions that can help. There are so many reasons not to use GPUs. By the time anyone seriously considers using GPUs they're, in my imagination, typically seriously starved for performance, and looking to control as much of the execution details as possible. GPU programmers don't want an automagic solution.
So I think the demand for easy GPU interfaces is just very weak, and therefore no effort has taken off. The amount of work needed to make it as easy to use as CPUs is massive, and the only reason anyone would even attempt to take this on is to lock you in to expensive hardware (see CUDA).
For a practical suggestion, have you taken a look at https://arrayfire.com/ ? It can run on both CUDA and OpenCL, and it has C++, Rust and Python bindings.
cuML - RAPIDS Machine Learning LibraryProject mention: Is it possible to run Sklearn models on a GPU? | /r/datascience | 2023-03-05
sklearn can't, bit take a look at cuML (https://github.com/rapidsai/cuml ). It uses the same API as sklearn but executes on GPU.
CUDA Templates for Linear Algebra SubroutinesProject mention: Optimization Techniques for GPU Programming [pdf] | news.ycombinator.com | 2023-08-09
I would recommend the course from Oxford (https://people.maths.ox.ac.uk/gilesm/cuda/). Also explore the tutorial section of cutlass (https://github.com/NVIDIA/cutlass/blob/main/media/docs/cute/...) if you want to learn more about high performance gemm.
Free Global Payroll designed for tech teams. Building a great tech team takes more than a paycheck. Zero payroll costs, get AI-driven insights to retain best talent, and delight them with amazing local benefits. 100% free and compliant.
HIP: C++ Heterogeneous-Compute Interface for PortabilityProject mention: Nvidia on the Mountaintop | news.ycombinator.com | 2023-08-28
AMD's equivalent is HIP , for sufficiently flexible definitions of "equivalent". I can't speak to how complete/correct/performant it is (I'm just a guy running tutorial/toy-level ML stuff on an RDNA1 card), but part of AMD's problem is that it might not practically matter how well they do this because the broader ecosystem support specifically for the CUDA stack is so entrenched.
ALIEN is a CUDA-powered artificial life simulation program.
LightSeq: A High Performance Library for Sequence Processing and Generation
Lightning fast C++/CUDA neural network frameworkProject mention: [D] Have their been any attempts to create a programming language specifically for machine learning? | /r/MachineLearning | 2023-02-11
In the opposite direction from your question is a very interesting project, TinyNN all implemented as close to the metal as possible and very fast: https://github.com/NVlabs/tiny-cuda-nn
HeavyDB (formerly OmniSciDB)
The C++ Standard Library for your entire system.
Fast inference engine for Transformer modelsProject mention: Faster Whisper Transcription with CTranslate2 | news.ycombinator.com | 2023-07-20
The original Whisper implementation from OpenAI uses the PyTorch deep learning framework. On the other hand, faster-whisper is implemented using CTranslate2  which is a custom inference engine for Transformer models. So basically it is running the same model but using another backend, which is specifically optimized for inference workloads.
CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.
CUDA on Intel GPUsProject mention: Why Nvidia Keeps Winning: The Rise of an AI Giant | news.ycombinator.com | 2023-07-06
> I don't think you understand just how insanely difficult it is to break into that market.
You're right, I have no clue nor have I ever tried myself.
> Even with apple money or something like that, it's a losing prospect because in the time it'll take you to get up and off the ground (which is FOREVER) your competition will crush you.
This I find hard to believe, do you have a source or reference for that claim? Companies with that amount of cash are hardly going to be crushed by competition be it direct or indirect. Anyway, I'm talking more about the Intels and AMDs of this world.
We have very lacklustre efforts from players I won't name with their Zluda library (https://github.com/vosen/ZLUDA) which I got REALLY excited about, until I read the README.txt. Four contributors, last commit early 2021.
Why, oh why, is it this bad?
Vulkan/CUDA/HIP/OpenCL/Level Zero/Metal Fast Fourier Transform libraryProject mention: [P] - VkFFT now supports quad precision (double-double) FFT computation on GPU | /r/MachineLearning | 2023-09-27
Hello, I am the creator of the VkFFT - GPU Fast Fourier Transform library for Vulkan/CUDA/HIP/OpenCL/Level Zero and Metal. In the latest update, I have added support for quad-precision double-double emulation for FFT calculation on most modern GPUs. I understand that modern ML is going in the opposite low-precision direction, but I still think that it may be useful to have this functionality at least for some prototyping and development of concepts.
Making it easier to work with shaders
Fast Neural Machine Translation in C++Project mention: [P] A CLI tool for easy transformer sequence classifier training and inference | /r/MachineLearning | 2023-02-01
As a reference, I forked https://github.com/marian-nmt/marian privately to support sequence tagging tasks. With a positional loss mask, It can also support sequence classificaiton.
stdgpu: Efficient STL-like Data Structures on the GPU
Implementation of SYCL and C++ standard parallelism for CPUs and GPUs from all vendors: The independent, community-driven compiler for C++-based heterogeneous programming models. Lets applications adapt themselves to all the hardware in the system - even at runtime!Project mention: Offloading standard C++ PSTL to Intel, NVIDIA and AMD GPUs with AdaptiveCpp | /r/cpp | 2023-09-24
AdaptiveCpp (formerly known as hipSYCL) is an independent, open source, clang-based heterogeneous C++ compiler project. I thought some of you might be interested in knowing that we recently added support to offload standard C++ parallel STL algorithms to GPUs from all major vendors. E.g.:
An efficient C++17 GPU numerical computing library with Python-like syntax
VUDA is a header-only library based on Vulkan that provides a CUDA Runtime API interface for writing GPU-accelerated applications.Project mention: VUDA: A Vulkan Implementation of CUDA | /r/patient_hackernews | 2023-07-01
Analyze your C and C++ projects with just one click.. SonarCloud, a cloud-based static analysis tool for your CI/CD workflows, offers a one-click automatic analysis of C and C++ projects hosted on GitHub. Zero configuration and free for open-source projects! Analyze free.
C++ Cuda related posts
Offloading standard C++ PSTL to Intel, NVIDIA and AMD GPUs with AdaptiveCpp
1 project | /r/cpp | 24 Sep 2023
Am I not good enough?
1 project | /r/learnprogramming | 23 Sep 2023
Alien v4.0 (Simulation)
2 projects | news.ycombinator.com | 1 Sep 2023
Optimization Techniques for GPU Programming [pdf]
1 project | news.ycombinator.com | 9 Aug 2023
Why Nvidia Keeps Winning: The Rise of an AI Giant
3 projects | news.ycombinator.com | 6 Jul 2023
VUDA: A Vulkan Implementation of CUDA
1 project | /r/patient_hackernews | 1 Jul 2023
VUDA: A Vulkan Implementation of CUDA
3 projects | news.ycombinator.com | 1 Jul 2023
A note from our sponsor - Onboard AI
getonboard.dev | 1 Oct 2023
What are some of the best open-source Cuda projects in C++? This list will help you: