batched-fn
tch-rs
batched-fn | tch-rs | |
---|---|---|
1 | 37 | |
17 | 3,860 | |
- | - | |
3.7 | 7.5 | |
about 2 months ago | 4 days ago | |
Rust | Rust | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
batched-fn
-
Processing a batch of requests for deep learning inference on a rust server
Some research: - Found a crate for exactly what I want, called batched_fn which seems to do exactly what I want with a catch that I cannot run async tasks (download, preprocess, etc) within the batch handler, ie, it's specifically for inference. I've opened an issue about it. - What I plan to do, is : - The response handlers pass their id's to a batching mechanism, and have a reciver for the output channel(details below) - to have a batching mechanism that batches up image id's on high load. - Pass it to another thread that downloads, preprocesses it and infers form it - This thread passes it to the result channel that every response handler has a reciever for. Every response handler checks if the message that it's reciving is for itself, and accordingly returns a JSON API response
tch-rs
- Tch-Rs
-
Llama2.rs: One-file Rust implementation of Llama2
I wanted to do something like this but then I would miss on proper CUDA acceleration and lose performance compared to using torchlib.
I wrote a forgettable llama implementation for https://github.com/LaurentMazare/tch-rs (pytorch's torchlib rust binding).
-
Playing Atari Games in OCaml
I first encountered OCaml's PyTorch bindings because apparently they generate a C wrapper around PyTorch's C++ API, and Rust's PyTorch bindings use OCaml's C wrapper. See: https://github.com/LaurentMazare/tch-rs
-
llm: a Rust crate/CLI for CPU inference of LLMs, including LLaMA, GPT-NeoX, GPT-J and more
You could try looking at the min-GPT example of tch-rs. I'd also strongly suggest watching Karpathy's video to understand what's going on.
-
Simply explained: How does GPT work?
If you pefer to see it in code there's a succint gpt implementation here https://github.com/LaurentMazare/tch-rs/blob/main/examples/m...
-
Will I ever need python again if I learn rust other than for AI stuff?
Rust is fully compatible w/ C bindings, so even Python libraries written in C can be easily set up to work in Rust (and have been). For example, see PyTorch Rust bindings, which actually works faster than in Python because all of the glue code around the C++ API is in Rust instead of Python.
-
A Rust client library for interacting with Microsoft Airsim https://github.com/Sollimann/airsim-client
Pytorch
- [D] HuggingFace in Julia or Rust ?
- This year I tried solving AoC using Rust, here are my impressions coming from Python!
-
[Help Needed] Deployment of torchscript using rust
I have looked into this a bit and found some crates which help in loading torchscript models called tch-rs
What are some alternatives?
PERSIA - High performance distributed framework for training deep learning recommendation models based on PyTorch.
onnxruntime - ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
rust - Empowering everyone to build reliable and efficient software.
candle - Minimalist ML framework for Rust
zebra - Zcash - Financial Privacy in Rust 🦓
cbindgen - A project for generating C bindings from Rust code
wtpsplit - Code for Where's the Point? Self-Supervised Multilingual Punctuation-Agnostic Sentence Segmentation
veloren - An open world, open source voxel RPG inspired by Dwarf Fortress and Cube World. This repository is a mirror. Please submit all PRs and issues on our GitLab page.
rustlearn - Machine learning crate for Rust
burn - Burn is a new comprehensive dynamic Deep Learning Framework built using Rust with extreme flexibility, compute efficiency and portability as its primary goals. [Moved to: https://github.com/Tracel-AI/burn]
linfa - A Rust machine learning framework.
tokenizers - 💥 Fast State-of-the-Art Tokenizers optimized for Research and Production