llama.onnx
motorhead
llama.onnx | motorhead | |
---|---|---|
2 | 10 | |
324 | 829 | |
- | 1.3% | |
7.3 | 7.7 | |
10 months ago | 11 days ago | |
Python | Rust | |
GNU General Public License v3.0 only | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
llama.onnx
-
Qnap TS-264
You can find LLM models in the onnx format here: https://github.com/tpoisonooo/llama.onnx
-
Langchain question and answer without openai
You also need a LLM to do this. Please check this out to pick one up from the llama family. Other works like llama.onnx, alpaca-native and llama model on hugging face are also worth checking.
motorhead
- Motorhead is a memory and information retrieval server for LLMs
-
Comparison of Vector Databases
Metal [1] is another one on my radar. Their API looks super simple.
Disclosures: None
[1] https://getmetal.io
-
Any Alternatives to Langchain?
Any alternatives? I found this Rust based project that might be interesting: https://github.com/getmetal/motorhead
- RasaGPT: First headless LLM chatbot built on top of Rasa, Langchain and FastAPI
-
Langchain question and answer without openai
you could run motorhead on docker https://github.com/getmetal/motorhead
-
How to use Enum with Vec to parse the mixed data vector from RedisSearch
The code is found using GitHub search FT.SEARCH inside https://github.com/getmetal/motorhead/blob/main/src/models.rs and adapted.
-
Memory in production
All the examples that Langchain gives are for persisting memory locally which won't work in a serverless (statelesss) environment, and the one solution documented for stateless applications, getmetal/motorhead, is a containerized, Rust-based service we would have to run ourselves.
- Show HN: Motörhead, LLM Memory Server Built in Rust
-
OpenAI Embeddings API alternative?
I've only just signed up and haven't had a chance to build anything with it yet, but this might be something to consider https://getmetal.io/
- Motörhead – memory and information retrieval server for LLMs
What are some alternatives?
llama.cpp - LLM inference in C/C++
lmql - A language for constraint-guided and efficient LLM programming.
Chinese-LLaMA-Alpaca - ä¸æ–‡LLaMA&Alpaca大è¯è¨€æ¨¡åž‹+本地CPU/GPUè®ç»ƒéƒ¨ç½² (Chinese LLaMA & Alpaca LLMs)
NeMo-Guardrails - NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
fastT5 - âš¡ boost inference speed of T5 models by 5x & reduce the model size by 3x.
RasaGPT - 💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram
AST-1 - Join the movement led by IZX.ai to create the world's best open-source LLM.
kor - LLM(😽)
llama2.openvino - This sample shows how to implement a llama-based model with OpenVINO runtime
Abstract Feature Branch - abstract_feature_branch is a Ruby gem that provides a variation on the Branch by Abstraction Pattern by Paul Hammant and the Feature Toggles Pattern by Martin Fowler (aka Feature Flags) to enable Continuous Integration and Trunk-Based Development.
openvino - OpenVINOâ„¢ is an open-source toolkit for optimizing and deploying AI inference
rasa-haystack