text-embeddings-inference
llama-node
text-embeddings-inference | llama-node | |
---|---|---|
3 | 2 | |
2,073 | 849 | |
10.6% | 0.9% | |
8.9 | 8.6 | |
4 days ago | 10 months ago | |
Rust | Rust | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
text-embeddings-inference
-
HuggingFace text-generation-inference is reverting to Apache 2.0 License
Worth noting that this also impacts the great https://github.com/huggingface/text-embeddings-inference, which allows anyone to run state of the art embeddings with great performance.
- FLaNK Stack Weekly for 30 Oct 2023
- Fast inference for text models using Rust
llama-node
-
Tell HN: Rust Is the Superglue
You can practice your Rust skills by writing performant and/or gluey extensions for higher-level language such as NodeJS (checkout napi-rs) and Python or complementing JS in the browser if you target Webassembly.
For instance, checkout Llama-node https://github.com/Atome-FE/llama-node for an involved Rust-based NodeJS extension. Python has PyO3, a Rust-Python extension toolset: https://github.com/PyO3/pyo3.
They can help you leverage your Rust for writing cool new stuff.
-
Could I get a suggestion for a simple HTTP API with no GUI for llama.cpp?
Node.js: hlhr202/llama-node
What are some alternatives?
smartgpt - A program that provides LLMs with the ability to complete complex tasks using plugins.
ChainFury - 🦋 Production grade chaining engine behind TuneChat. Self host today!
auto-rust - auto-rust is an experimental project that automatically generate Rust code with LLM (Large Language Models) during compilation, utilizing procedural macros.
LLamaSharp - A C#/.NET library to run LLM models (🦙LLaMA/LLaVA) on your local device efficiently.
floneum - A toolkit for controllable, private AI on consumer hardware in rust
langchain-ask-pdf-local - An AI-app that allows you to upload a PDF and ask questions about it. It uses StableVicuna 13B and runs locally.
openv0 - AI generated UI components
llama-cpp.el - A client for llama-cpp server
CSGHub - CSGHub is an opensource large model assets platform just like on-premise huggingface which helps to manage datasets, model files, codes and more. CSGHub是一个开源、可信的大模型资产管理平台,可帮助用户治理LLM和LLM应用生命周期中涉及到的资产(数据集、模型文件、代码等)。CSGHub提供类似私有化的Huggingface功能,以类似OpenStack Glance管理虚拟机镜像、Harbor管理容器镜像以及Sonatype Nexus管理制品的方式,实现对LLM资产的管理。欢迎关注反馈和Star⭐️
gpt4all.unity - Bindings of gpt4all language models for Unity3d running on your local machine
anansi - open source tooling for AI search and understanding
langtorch - 🔥 Building composable LLM applications & workflow with Java.