SaaSHub helps you find the best software and product alternatives Learn more →
Top 10 C++ llama Projects
-
LocalAI
:robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
cortex
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan (by janhq)
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
llama-server-chat-terminal-client
Lightweight chat terminal-interface for llama.cpp server compilable for windows and linux.
Project mention: The prospects for 128 bit processors (John R. Mashey, 1995) | news.ycombinator.com | 2024-05-21You don't need 128 bits for memory addressing, but for just processing - yes, and in fact 128 bits is far less than we're using already! If you look at https://github.com/ggerganov/llama.cpp you'll see this line:
> AVX, AVX2 and AVX512 support for x86 architectures
Guess what the 512 in AVX512 stands for?;)
On GPUs I'm pretty sure the same thing is in play, but I'm less familiar. A quick search turns up ex. https://developer.nvidia.com/blog/implementing-high-precisio... which makes me think yes.
Project mention: LocalAI: Self-hosted OpenAI alternative reaches 2.14.0 | news.ycombinator.com | 2024-05-03
Jan incorporates a lightweight, built-in inference server called Nitro. Nitro supports both llama.cpp and NVIDIA's TensorRT-LLM engines. This means many open LLMs in the GGUF format are supported. Jan's Model Hub is designed for easy installation of pre-configured models but it also allows you to install virtually any model from Hugging Face or even your own.
Project mention: New to this community, most models I download fail and end up in a core dump | /r/LocalLLaMA | 2023-05-23If you want to use that model specifically, check out : https://github.com/kuvaus/LlamaGPTJ-chat
Python sits on the C-glue segment of programming languages (where Perl, PHP, Ruby and Node are also notable members). Being a glue language means having APIs to a lot of external toolchains written in not only C/C++ but many other compiled languages, APIs and system resources. Conda, virtualenv, etc. are godsend modules for making it all work, or even better, to freeze things once they all work, without resourcing to Docker, VMs or shell scripts. It's meant for application and DevOps people who need to slap together, ie, ML, Numpy, Elasticsearch, AWS APIs and REST endpoints and Get $hit Done.
It's annoying to see them "glueys" compared to the binary compiled segment where the heavy lifting is done. Python and others exist to latch on and assimilate. Resistance is futile:
https://pypi.org/project/pyllamacpp/
https://www.npmjs.com/package/llama-node
https://packagist.org/packages/kambo/llama-cpp-php
https://github.com/yoshoku/llama_cpp.rb
Project mention: Show HN: Collider – the platform for local LLM debug and inference at warp speed | news.ycombinator.com | 2023-11-30
C++ llama related posts
-
The prospects for 128 bit processors (John R. Mashey, 1995)
-
New exponent functions that make SiLU and SoftMax 2x faster, at full acc
-
Gemini Flash
-
Ggml: Add Flash Attention
-
Structured: Extract Data from Unstructured Input with LLM
-
IBM Granite: A Family of Open Foundation Models for Code Intelligence
-
Ask HN: Affordable hardware for running local large language models?
-
A note from our sponsor - SaaSHub
www.saashub.com | 21 May 2024
Index
What are some of the best open-source llama projects in C++? This list will help you:
Project | Stars | |
---|---|---|
1 | llama.cpp | 58,425 |
2 | LocalAI | 20,346 |
3 | PowerInfer | 7,008 |
4 | cortex | 1,635 |
5 | LlamaGPTJ-chat | 211 |
6 | llama_cpp.rb | 144 |
7 | booster | 119 |
8 | pyllamacpp | 59 |
9 | llama-server-chat-terminal-client | 10 |
10 | llama-chat | 7 |
Sponsored