llm
crates.io
llm | crates.io | |
---|---|---|
41 | 662 | |
5,954 | 2,814 | |
3.1% | 1.6% | |
9.4 | 10.0 | |
about 2 months ago | 6 days ago | |
Rust | Rust | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
llm
-
Open-sourcing a simple automation/agent workflow builder
We're open-sourcing a project that lets you build simple automations/agent workflows that use LLMs for different tasks. Kinda like Zapier or IFTTT but focused on using natural language to accomplish your tasks.It's super early but we'd love to start getting feedback to steer it in the right direction. It currently supports OpenAI and local models through llm.
-
Meta's Segment Anything written with C++ / GGML
> Tensorflow is a C++ framework that has Python bindings and a Python library, but when the models are served they are running on C++
Sure, and it's only a simple 20 step process that involves building Tensorflow from source. Yeay!
https://medium.com/@hamedmp/exporting-trained-tensorflow-mod...
Let me see what the process for compiling a LLM written in Rust is....
https://github.com/rustformers/llm
cargo install llm-cli
-
Announcing Floneum (A open source graph editor for local AI workflows written in rust)
Floneum is a graph editor for local AI workflows. It uses llm to run large language models locally, egui, and dioxus for the frontend, and wasmtime for the plugin system. If you are interested in the project, consider joining the discord, or building a plugin for Floneum in rust using WASI
- are there anytools or frameworks similar to "langchain" or "llamaindexbut implemented or designed in a language other than python?
-
(1/2) May 2023
Run inference for Large Language Models on CPU, with Rust (https://github.com/rustformers/llm)
-
I built a multi-platform desktop app to easily download and run models, open source btw
On the rustformers github page I see that one of the commands to generate the answer is llm llama infer -m ggml-gpt4all-j-v1.3-groovy.bin -p "Rust is a cool programming language because", my basic idea for now is to change the Tauri app to let it do -p prompt, which receives from my code through the link or through a shared variable (if I don't use the link and start different times your app)
- Weekly Megathread - 14 May 2023
-
rustformers/llm: Run inference for Large Language Models on CPU, with Rust π¦ππ¦
wonnx has done some fantastic work in this regard, so that's where we plan to start once we get there. In terms of general discussion of alternate backends, see this issue.
- llm: a Rust crate/CLI for CPU inference of LLMs, including LLaMA, GPT-NeoX, GPT-J and more
crates.io
-
Create a Custom GitHub Action in Rust
Rust has a rich ecosystem of frameworks and libraries that let you read, parse, and manipulate text files, interact with cloud services and databases, and perform any other job that your project's development workflow may require. And because of its strong typing and tight memory management, you are much less likely to write programs that behave unexpectedly in production.
-
Rust Keyword Extraction: Creating the YAKE! algorithm from scratch
All the code discussed in this article can be accessed through this repository. For integration with existing projects consider using keyword_extraction crate available on crates.io.
-
Migrating a JavaScript frontend to Leptos, a Rust framework
So, be sure to double-check your critical libraries and be sure their alternatives exist in the Rust ecosystem. Thereβs a good chance the crates you need are available in Rust's crates.io repository.
-
Learning Rust: A clean start
The previous section was very simple, this section is also very simple but introduces us to cargo which is Rust's package manager, as a JS dev my mind goes straight to NPM.
-
#2 Rust - Cargo Package Manager
Now, there has to be a place where all these packages come from. Similar to npmjs registry, where all node packages are registered, stored and retrieved, Rust also has something called crates.io where many helpful packages and dependencies are registered.
-
Rust π¦ Installation + Hello World
Before proceeding, let's check https://crates.io/, the official Rust package registry.
-
Underestimating rust for my Project.
The most thrilling aspect has been the joy of writing the backend. It's like every struct, enum, and method in Rust forms this interconnected Multiverse of code , which you can see in crates.io which is best Documentation experience I Ever Had.
-
Top 10 Rusty Repositories for you to start your Open Source Journey
5. Crates.io
-
Project Structure Clarification Coming From Python - With Example
When using crates from eg. crates.io, and also things like std and core
-
Cargo has never frustrated me like npm or pip has. Does Cargo ever get frustrating? Does anyone ever find themselves in dependency hell?
Vendoring your packages was very tedious to even remotely get to work with Cargo. I spent a very long time getting Cargo to work together with cargo-local-registry. We vendor crates from crates.io and a custom internal registry.
What are some alternatives?
llama.cpp - LLM inference in C/C++
docs.rs - crates.io documentation generator
ggml - Tensor library for machine learning
plotters - A rust drawing library for high quality data plotting for both WASM and native, statically and realtimely π¦ ππ
GPTQ-for-LLaMa - 4 bits quantization of LLaMA using GPTQ
Cargo - The Rust package manager
alpaca-lora - Instruct-tune LLaMA on consumer hardware
trunk - Build, bundle & ship your Rust WASM application to the web.
alpaca.cpp - Locally run an Instruction-Tuned Chat-Style LLM
gtk4-rs - Rust bindings of GTK 4
SD-CN-Animation - This script allows to automate video stylization task using StableDiffusion and ControlNet.
Rocket - A web framework for Rust.