I made an Open-Source frontend for the Koala LLM using llama.cpp and llama-cpp-python

This page summarizes the projects mentioned and recommended in the original post on /r/opensource

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • llama.cpp

    LLM inference in C/C++

  • There is this language model called Koala, developed by researchers at UC Berkeley. It's based on LLaMA by Meta. Using llama.cpp, one can easily run this model on the CPU instead on a dedicated GPU. With the python library llama-cpp-python, it can be used with Python + it has an API-Server included.

  • llama-cpp-python

    Python bindings for llama.cpp

  • There is this language model called Koala, developed by researchers at UC Berkeley. It's based on LLaMA by Meta. Using llama.cpp, one can easily run this model on the CPU instead on a dedicated GPU. With the python library llama-cpp-python, it can be used with Python + it has an API-Server included.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • Eucalyptus-Chat

    Discontinued A frontend for large language models like 🐨 Koala or 🦙 Vicuna running on CPU with llama.cpp, using the API server library provided by llama-cpp-python. NOTE: I had to discontinue this project because its maintenance takes more time than I can and want to invest. Feel free to fork :)

  • I put this all together and built a chat-like frontend on top of it: Eucalyptus. It's not well documented, but I think you should be able to use it without problems - if you have the model files for Koala (that must be converted & should be quantized using llama.cpp by the way).

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts