motorhead
llama.onnx
Our great sponsors
motorhead | llama.onnx | |
---|---|---|
10 | 2 | |
822 | 323 | |
2.6% | - | |
8.0 | 7.3 | |
9 days ago | 10 months ago | |
Rust | Python | |
Apache License 2.0 | GNU General Public License v3.0 only |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
motorhead
- Motorhead is a memory and information retrieval server for LLMs
-
Comparison of Vector Databases
Metal [1] is another one on my radar. Their API looks super simple.
Disclosures: None
[1] https://getmetal.io
-
Any Alternatives to Langchain?
Any alternatives? I found this Rust based project that might be interesting: https://github.com/getmetal/motorhead
- RasaGPT: First headless LLM chatbot built on top of Rasa, Langchain and FastAPI
-
Langchain question and answer without openai
you could run motorhead on docker https://github.com/getmetal/motorhead
-
How to use Enum with Vec to parse the mixed data vector from RedisSearch
The code is found using GitHub search FT.SEARCH inside https://github.com/getmetal/motorhead/blob/main/src/models.rs and adapted.
-
Memory in production
All the examples that Langchain gives are for persisting memory locally which won't work in a serverless (statelesss) environment, and the one solution documented for stateless applications, getmetal/motorhead, is a containerized, Rust-based service we would have to run ourselves.
- Show HN: Motörhead, LLM Memory Server Built in Rust
-
OpenAI Embeddings API alternative?
I've only just signed up and haven't had a chance to build anything with it yet, but this might be something to consider https://getmetal.io/
- Motörhead – memory and information retrieval server for LLMs
llama.onnx
-
Qnap TS-264
You can find LLM models in the onnx format here: https://github.com/tpoisonooo/llama.onnx
-
Langchain question and answer without openai
You also need a LLM to do this. Please check this out to pick one up from the llama family. Other works like llama.onnx, alpaca-native and llama model on hugging face are also worth checking.
What are some alternatives?
lmql - A language for constraint-guided and efficient LLM programming.
llama.cpp - LLM inference in C/C++
NeMo-Guardrails - NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
Chinese-LLaMA-Alpaca - ä¸æ–‡LLaMA&Alpaca大è¯è¨€æ¨¡åž‹+本地CPU/GPUè®ç»ƒéƒ¨ç½² (Chinese LLaMA & Alpaca LLMs)
RasaGPT - 💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram
fastT5 - âš¡ boost inference speed of T5 models by 5x & reduce the model size by 3x.
kor - LLM(😽)
AST-1 - Join the movement led by IZX.ai to create the world's best open-source LLM.
Abstract Feature Branch - abstract_feature_branch is a Ruby gem that provides a variation on the Branch by Abstraction Pattern by Paul Hammant and the Feature Toggles Pattern by Martin Fowler (aka Feature Flags) to enable Continuous Integration and Trunk-Based Development.
llama2.openvino - This sample shows how to implement a llama-based model with OpenVINO runtime
rasa-haystack
openvino - OpenVINOâ„¢ is an open-source toolkit for optimizing and deploying AI inference