llama2.rs
tch-rs
llama2.rs | tch-rs | |
---|---|---|
3 | 37 | |
981 | 3,899 | |
- | - | |
8.9 | 7.5 | |
6 months ago | 19 days ago | |
Rust | Rust | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
llama2.rs
-
Ask HN: Cheapest hardware to run Llama 2 70B
This code runs Llama2 quantized and unquantized in a roughly minimal way: https://github.com/srush/llama2.rs (though extracting the quantized 70B weights takes a lot of RAM). I'm running the 13B quantized model on ~10-11GB of CPU memory.
-
Candle: Torch Replacement in Rust
Nowhere near as neat as candle or ggml, but just released a 4-bit rust llama2 implementation with simd. Runs pretty fast.
https://github.com/srush/llama2.rs/
- Llama2.rs: One-file Rust implementation of Llama2
tch-rs
- Tch-Rs
-
Llama2.rs: One-file Rust implementation of Llama2
I wanted to do something like this but then I would miss on proper CUDA acceleration and lose performance compared to using torchlib.
I wrote a forgettable llama implementation for https://github.com/LaurentMazare/tch-rs (pytorch's torchlib rust binding).
-
Playing Atari Games in OCaml
I first encountered OCaml's PyTorch bindings because apparently they generate a C wrapper around PyTorch's C++ API, and Rust's PyTorch bindings use OCaml's C wrapper. See: https://github.com/LaurentMazare/tch-rs
-
llm: a Rust crate/CLI for CPU inference of LLMs, including LLaMA, GPT-NeoX, GPT-J and more
You could try looking at the min-GPT example of tch-rs. I'd also strongly suggest watching Karpathy's video to understand what's going on.
-
Simply explained: How does GPT work?
If you pefer to see it in code there's a succint gpt implementation here https://github.com/LaurentMazare/tch-rs/blob/main/examples/m...
-
Will I ever need python again if I learn rust other than for AI stuff?
Rust is fully compatible w/ C bindings, so even Python libraries written in C can be easily set up to work in Rust (and have been). For example, see PyTorch Rust bindings, which actually works faster than in Python because all of the glue code around the C++ API is in Rust instead of Python.
-
A Rust client library for interacting with Microsoft Airsim https://github.com/Sollimann/airsim-client
Pytorch
- [D] HuggingFace in Julia or Rust ?
- This year I tried solving AoC using Rust, here are my impressions coming from Python!
-
[Help Needed] Deployment of torchscript using rust
I have looked into this a bit and found some crates which help in loading torchscript models called tch-rs
What are some alternatives?
burn - Burn is a new comprehensive dynamic Deep Learning Framework built using Rust with extreme flexibility, compute efficiency and portability as its primary goals.
onnxruntime - ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
candle - Minimalist ML framework for Rust
euclid - Geometry primitives (basic linear algebra) for Rust
cbindgen - A project for generating C bindings from Rust code
exllama - A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
wtpsplit - Code for Where's the Point? Self-Supervised Multilingual Punctuation-Agnostic Sentence Segmentation
llama.cpp - LLM inference in C/C++
veloren - An open world, open source voxel RPG inspired by Dwarf Fortress and Cube World. This repository is a mirror. Please submit all PRs and issues on our GitLab page.
petals - 🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
rustlearn - Machine learning crate for Rust