Our great sponsors
-
LocalAI
:robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
-
gpt-discord-bot
Example Discord bot written in Python that uses the completions API to have conversations with the `text-davinci-003` model, and the moderations API to filter the messages.
-
turbopilot
Discontinued Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
serge
A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
-
web-llm
Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.
check out the examples, you can plug in chatbot-ui! https://github.com/go-skynet/LocalAI/tree/master/examples/chatbot-ui
https://chat.lmsys.org/ now has a "chatbot arena" where you can pick two models and see their simultaneous responses to the same prompt. The demo service they're using is open source (https://github.com/lm-sys/FastChat ) and some of the models they're using are also open source, but the majority of them are patches on top of the leaked Meta llama one and thus are of questionable licensing
Some of them are available I'm huggingface, you can search for "ggml". I've listed in the README few of the most common ones like https://github.com/nomic-ai/gpt4all . I'm working as well to simplify getting free licensed models in a more maintainable way, and ease out that in the API, stay tuned!
How does this compare to https://github.com/oobabooga/text-generation-webui ?
Yeah, looks like there might be some current/future support for some of these LLMs with amd; https://github.com/RadeonOpenCompute/ROCm/discussions/1836 It’s just not as robust for what I understand unfortunately.
try the instructions on this github repo https://github.com/antimatter15/alpaca.cpp, its not the best one but I was able to run this model on my linux machine with 16GB memory, I think its a good starting point.
Thanks for sharing your hard work. How would you say LocalAI differs from Serge?
Now that WebGPU is coming in Chrome 113, I am hoping to see more "in the browser" LLM's, like the amazing demo from MLC AI https://mlc.ai/web-llm/
Related posts
- Best way to use AMD CPU and GPU
- Meet Atom the GPT Assistant, an AI-powered Smart Home Assistant. It's like Google Assistant but with endless possibility of ChatGPT, it's like Siri but with extensibility of Open Source power.
- chatgpt alternative
- Show HN: LlamaGPT – Self-hosted, offline, private AI chatbot, powered by Llama 2
- LeCun: Qualcomm working with Meta to run Llama-2 on mobile devices