Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Top 11 C++ llm Projects
-
LocalAI
:robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
nitro
An inference server on top of llama.cpp. OpenAI-compatible API, queue, & scaling. Embed a prod-ready, local inference engine in your apps. Powers Jan (by janhq)
-
distributed-llama
Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
redis-llm
redis-llm integrates LLM with Redis. It can help LLM to access private data, and remember long chat history.
well https://github.com/ggerganov/llama.cpp/issues/6849
I'd like to see a comparison to nitro https://github.com/janhq/nitro which has been fantastic for running a local LLM.
There's https://github.com/saharNooby/rwkv.cpp, which related-ish[0] to ggml/llama.cpp
[0]: https://github.com/ggerganov/llama.cpp/issues/846
What is your novelty? I just see HNSW.
What is the relationship between awa and vearch? You both have a 'gamma' index with identical copy:
https://github.com/awa-ai/awadb/tree/main/awadb/db_engine/in...
Python sits on the C-glue segment of programming languages (where Perl, PHP, Ruby and Node are also notable members). Being a glue language means having APIs to a lot of external toolchains written in not only C/C++ but many other compiled languages, APIs and system resources. Conda, virtualenv, etc. are godsend modules for making it all work, or even better, to freeze things once they all work, without resourcing to Docker, VMs or shell scripts. It's meant for application and DevOps people who need to slap together, ie, ML, Numpy, Elasticsearch, AWS APIs and REST endpoints and Get $hit Done.
It's annoying to see them "glueys" compared to the binary compiled segment where the heavy lifting is done. Python and others exist to latch on and assimilate. Resistance is futile:
https://pypi.org/project/pyllamacpp/
https://www.npmjs.com/package/llama-node
https://packagist.org/packages/kambo/llama-cpp-php
https://github.com/yoshoku/llama_cpp.rb
Project mention: Show HN: Collider – the platform for local LLM debug and inference at warp speed | news.ycombinator.com | 2023-11-30
Project mention: Show HN: Redis-LLM – Redis module integrates LLM with Redis | news.ycombinator.com | 2023-07-10
C++ llm related posts
- Any Online Communities on Local/Home AI?
- Phi-3 Weights Released
- Llama.cpp Working on Support for Llama3
- Mixtral 8x22B
- Llama.cpp: Improve CPU prompt eval speed
- Ollama 0.1.32: WizardLM 2, Mixtral 8x22B, macOS CPU/GPU model split
- Distributed Grok-1 (314B)
-
A note from our sponsor - InfluxDB
www.influxdata.com | 28 Apr 2024
Index
What are some of the best open-source llm projects in C++? This list will help you:
Project | Stars | |
---|---|---|
1 | llama.cpp | 56,891 |
2 | LocalAI | 19,593 |
3 | PowerInfer | 6,942 |
4 | koboldcpp | 3,749 |
5 | nitro | 1,572 |
6 | rwkv.cpp | 1,097 |
7 | distributed-llama | 734 |
8 | awadb | 158 |
9 | llama_cpp.rb | 129 |
10 | collider | 117 |
11 | redis-llm | 44 |
Sponsored