Our great sponsors
-
LocalAI
:robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
you could run motorhead on docker https://github.com/getmetal/motorhead
Maybe llama.cpp is what you might need. It doesn't even need a GPU and can run on mobile device.
You also need a LLM to do this. Please check this out to pick one up from the llama family. Other works like llama.onnx, alpaca-native and llama model on hugging face are also worth checking.
Related posts
- Motorhead is a memory and information retrieval server for LLMs
- PostgresML
- [P] pgml-chat: A command-line tool for deploying low-latency knowledge-based chatbots
- Python SDK for PostgresML with scalable LLM embedding memory and text generation
- [P] Python SDK for PostgresML w/ scalable LLM embedding memory and text generation