burn
Rust-CUDA
burn | Rust-CUDA | |
---|---|---|
34 | 37 | |
4,845 | 2,894 | |
- | 2.4% | |
8.9 | 0.0 | |
6 months ago | 6 months ago | |
Rust | Rust | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
burn
-
Burn 0.10.0 Released 🔥 (Deep Learning Framework)
Release Note: https://github.com/burn-rs/burn/releases/tag/v0.10.0
- Deep Learning Framework in Rust: Burn 0.10.0 Released
-
Why Rust Is the Optimal Choice for Deep Learning, and How to Start Your Journey with the Burn Deep Learning Framework
The comprehensive, open-source deep learning framework in Rust, Burn, has recently undergone significant advancements in its latest release, highlighted by the addition of The Burn Book 🔥. There has never been a better moment to embark on your deep learning journey with Rust, as this book will guide you through your initial project, providing extensive explanations and links to relevant resources.
-
Candle: Torch Replacement in Rust
Burn (deep learning framework in rust) has WGPU backend (WebGPU) already. Check it out https://github.com/burn-rs/burn. It was released recently.
- Burn – A Flexible and Comprehensive Deep Learning Framework in Rust
-
Announcing Burn-Wgpu: New Deep Learning Cross-Platform GPU Backend
For more details about the latest release see the release notes: https://github.com/burn-rs/burn/releases/tag/v0.8.0.
-
Are there any ML crates that would compile to WASM?
Tract is the most well known ML crate in Rust, which I believe can compile to WASM - https://github.com/sonos/tract/. Burn may also be useful - https://github.com/burn-rs/burn.
-
Any working wgpu compute example that would run in a browser?
We, the burn team, are working on the wgpu backend (WebGPU) for Burn deep learning framework. You can check out the current state: https://github.com/burn-rs/burn/tree/main/burn-wgpu
-
I’ve fallen in love with rust so now what?
Here is the project: https://github.com/burn-rs/burn
-
Is anyone doing Machine Learning in Rust?
Disclaimer, I'm the main author of Burn https://burn-rs.github.io.
Rust-CUDA
-
[Media] Anyone try writing a ray tracer with rust? It's pretty fun!
Source code [here](https://github.com/ihawn/RTracer) if anyone is interested in taking a look or giving feedback. As a side question, does anyone have any general advise on getting GPU compute working with rust? I tried [this project](https://github.com/Rust-GPU/Rust-CUDA) but had a bunch of issues (And it doesn't look like an active repo anyways)
-
Is rust or python better for Machine learning? Or is there enough decent frameworks?
You have this https://github.com/Rust-GPU/Rust-CUDA
-
toolchain nightly package building issue
What I'm trying to do is check out https://github.com/Rust-GPU/Rust-CUDA for a class project.
- [Rust] État de GPGPU en 2022
- Which crate for CUDA in Rust?
-
Announcing cudarc and fully GPU accelerated dfdx: ergonomic deep learning ENTIRELY in rust, now with CUDA support and tensors with mixed compile and runtime dimensions!
Be warned, NON_BLOCKING streams do not fully synchronize with sync host to device copies. They are not guaranteed to actually finish by the time they return. Meaning its possible to initiate a copy, then initiate a kernel launch, and have the copy be unfinished by the time the kernel is launched. This caused so many confusing bugs that i personally decided to stop using NON_BLOCKING altogether in rust-cuda. https://github.com/Rust-GPU/Rust-CUDA/issues/15
-
In which circumstances is C++ better than Rust?
- Cuda is not doing by FFI linking, instead is compiling CUDA code natively in Rust https://github.com/Rust-GPU/Rust-CUDA and even if it not complete as the C++ SDK is more than a toy
- I learned 7 programming languages so you don't have to
-
GNU Octave
Given your criteria, you might want to consider (modern) C++.
* Fast - in many cases faster than Rust, although the difference is inconsequential relative to Python-to-Rust improvement I guess.
* _Really_ utilize CUDA, OpenCL, Vulcan etc. Specifically, Rust GPU is limited in its supported features, see: https://github.com/Rust-GPU/Rust-CUDA/blob/master/guide/src/... ...
* Host-side use of CUDA is at least as nice, and probably nicer, than what you'll get with Rust. That is, provided you use my own Modern C++ wrappers for the CUDA APIs: https://github.com/eyalroz/cuda-api-wrappers/ :-) ... sorry for the shameless self-plug.
* ... which brings me to another point: Richer offering of libraries for various needs than Rust, for you to possibly utilize.
* Easier to share than Rust. A target system is less likely to have an appropriate version of Rust and the surrounding ecosystem.
There are downsides, of course, but I was just applying your criteria.
-
Your average rustafarians
Technically, yes. There are crates for OpenCL and CUDA, although official ROCm support does not exist yet.
What are some alternatives?
candle - Minimalist ML framework for Rust
rust-gpu - 🐉 Making Rust a first-class language and ecosystem for GPU shaders 🚧
dfdx - Deep learning in Rust, with shape checked tensors and neural networks
wgpu - Cross-platform, safe, pure-rust graphics api.
tch-rs - Rust bindings for the C++ api of PyTorch.
rust-ndarray - ndarray: an N-dimensional array with array views, multidimensional slicing, and efficient operations
Graphite - 2D raster & vector editor that melds traditional layers & tools with a modern node-based, non-destructive, procedural workflow.
CUDA.jl - CUDA programming in Julia.
tract - Tiny, no-nonsense, self-contained, Tensorflow and ONNX inference [Moved to: https://github.com/sonos/tract]
GLSL - GLSL Shading Language Issue Tracker
L2 - l2 is a fast, Pytorch-style Tensor+Autograd library written in Rust
WeasyPrint - The awesome document factory