burn
Burn is a new comprehensive dynamic Deep Learning Framework built using Rust with extreme flexibility, compute efficiency and portability as its primary goals. [Moved to: https://github.com/Tracel-AI/burn] (by burn-rs)
candle
Minimalist ML framework for Rust (by huggingface)
burn | candle | |
---|---|---|
34 | 20 | |
4,845 | 17,166 | |
- | 1.5% | |
8.9 | 9.5 | |
over 1 year ago | 6 days ago | |
Rust | Rust | |
Apache License 2.0 | Apache License 2.0 |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
burn
Posts with mentions or reviews of burn.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-08-08.
-
Burn 0.10.0 Released 🔥 (Deep Learning Framework)
Release Note: https://github.com/burn-rs/burn/releases/tag/v0.10.0
- Deep Learning Framework in Rust: Burn 0.10.0 Released
-
Why Rust Is the Optimal Choice for Deep Learning, and How to Start Your Journey with the Burn Deep Learning Framework
The comprehensive, open-source deep learning framework in Rust, Burn, has recently undergone significant advancements in its latest release, highlighted by the addition of The Burn Book 🔥. There has never been a better moment to embark on your deep learning journey with Rust, as this book will guide you through your initial project, providing extensive explanations and links to relevant resources.
-
Candle: Torch Replacement in Rust
Burn (deep learning framework in rust) has WGPU backend (WebGPU) already. Check it out https://github.com/burn-rs/burn. It was released recently.
- Burn – A Flexible and Comprehensive Deep Learning Framework in Rust
-
Announcing Burn-Wgpu: New Deep Learning Cross-Platform GPU Backend
For more details about the latest release see the release notes: https://github.com/burn-rs/burn/releases/tag/v0.8.0.
-
Are there any ML crates that would compile to WASM?
Tract is the most well known ML crate in Rust, which I believe can compile to WASM - https://github.com/sonos/tract/. Burn may also be useful - https://github.com/burn-rs/burn.
-
Any working wgpu compute example that would run in a browser?
We, the burn team, are working on the wgpu backend (WebGPU) for Burn deep learning framework. You can check out the current state: https://github.com/burn-rs/burn/tree/main/burn-wgpu
-
I’ve fallen in love with rust so now what?
Here is the project: https://github.com/burn-rs/burn
-
Is anyone doing Machine Learning in Rust?
Disclaimer, I'm the main author of Burn https://burn-rs.github.io.
candle
Posts with mentions or reviews of candle.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2025-05-03.
-
Conduit: A UI-less node-based system
I intend to grow this into an open-source project because deep inside, this is ideally how I would like ComfyUI to be. There's still a long journey ahead for building all the custom nodes, which is especially challenging given that the majority of code for AI workflows is written in Python. However, with my hands-on experience with Candle and Burn libraries, I may be able to get pretty close!
-
Lm.rs Minimal CPU LLM inference in Rust with no dependency
You are correct. This project is "on the CPU", so it will not utilize your GPU for computation. If you would like to try out a Rust framework that does support GPUs, [Candle](https://github.com/huggingface/candle/tree/main) may be worth exploring
-
Vector search in Manticore
While looking into how to create text embeddings quickly and directly, we discovered a few helpful tools that allowed us to achieve our goal. Consequently, we created an easy-to-use PHP extension that can generate text embeddings. This extension lets you pick any model from Sentence Transformers on HuggingFace. It is built on the CandleML framework, which is written in Rust and is a part of the well-known HuggingFace ecosystem. The PHP extension itself is also crafted in Rust using the php-ext-rs library. This approach ensures the extension runs fast while still being easy to develop.
-
karpathy/llm.c
Candle already exists[1], and it runs pretty well. Can use both CUDA and Metal backends (or just plain-old CPU).
[1] https://github.com/huggingface/candle
- Best alternative for python
-
Is there any LLM that can be installed with out python
Check out Candle! It's a Deep Learning framework for Rust. You can run LLMs in binaries.
-
Announcing Kalosm - an local first AI meta-framework for Rust
Kalosm is a meta-framework for AI written in Rust using candle. Kalosm supports local quantized large language models like Llama, Mistral, Phi-1.5, and Zephyr. It also supports other quantized models like Wuerstchen, Segment Anything, and Whisper. In addition to local models, Kalosm supports remote models like GPT-4 and ada embeddings.
-
RFC: candle-lora
I have been working on a machine learning library called candle-lora for Candle. It implementes a technique called LoRA (low rank adaptation), which allows you to reduce a model's trainable parameter count by wrapping and freezing old layers.
-
ExecuTorch: Enabling On-Device interference for embedded devices
[2] https://github.com/huggingface/candle/issues/313
-
[P] Open-source project to run locally LLMs in browser, such as Phi-1.5 for fully private inference
We provide full local inference in browser, by using libraries from Hugging Face like transformers.js or candle for WASM inference.
What are some alternatives?
When comparing burn and candle you can also consider the following projects:
tch-rs - Rust bindings for the C++ api of PyTorch.
Universal-G-Code-Sender - A cross-platform G-Code sender for GRBL, Smoothieware, TinyG and G2core.
Graphite - 2D vector & raster editor that melds traditional layers & tools with a modern node-based, non-destructive, procedural workflow.
gsender - Connect to and control grbl and grblHAL-based CNCs with ease
tract - Tiny, no-nonsense, self-contained, Tensorflow and ONNX inference [Moved to: https://github.com/sonos/tract]
bCNC - GRBL CNC command sender, autoleveler and g-code editor