C++ llm

Open-source C++ projects categorized as llm

Top 11 C++ llm Projects

  • llama.cpp

    LLM inference in C/C++

  • Project mention: Phi-3 Weights Released | news.ycombinator.com | 2024-04-23

    well https://github.com/ggerganov/llama.cpp/issues/6849

  • LocalAI

    :robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.

  • Project mention: Drop-In Replacement for ChatGPT API | news.ycombinator.com | 2024-01-24
  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • PowerInfer

    High-speed Large Language Model Serving on PCs with Consumer-grade GPUs

  • Project mention: FLaNK 25 December 2023 | dev.to | 2023-12-26
  • koboldcpp

    A simple one-file way to run various GGML and GGUF models with KoboldAI's UI

  • Project mention: Any Online Communities on Local/Home AI? | news.ycombinator.com | 2024-04-24
  • nitro

    An inference server on top of llama.cpp. OpenAI-compatible API, queue, & scaling. Embed a prod-ready, local inference engine in your apps. Powers Jan (by janhq)

  • Project mention: Ollama Python and JavaScript Libraries | news.ycombinator.com | 2024-01-24

    I'd like to see a comparison to nitro https://github.com/janhq/nitro which has been fantastic for running a local LLM.

  • rwkv.cpp

    INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model

  • Project mention: Eagle 7B: Soaring past Transformers | news.ycombinator.com | 2024-01-28

    There's https://github.com/saharNooby/rwkv.cpp, which related-ish[0] to ggml/llama.cpp

    [0]: https://github.com/ggerganov/llama.cpp/issues/846

  • distributed-llama

    Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.

  • Project mention: Distributed Grok-1 (314B) | news.ycombinator.com | 2024-04-15
  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • awadb

    AI Native database for embedding vectors

  • Project mention: An AI Native database for embedding vectors | news.ycombinator.com | 2023-09-10

    What is your novelty? I just see HNSW.

    What is the relationship between awa and vearch? You both have a 'gamma' index with identical copy:

    https://github.com/awa-ai/awadb/tree/main/awadb/db_engine/in...

  • llama_cpp.rb

    llama_cpp provides Ruby bindings for llama.cpp

  • Project mention: Llama.cpp: Full CUDA GPU Acceleration | news.ycombinator.com | 2023-06-12

    Python sits on the C-glue segment of programming languages (where Perl, PHP, Ruby and Node are also notable members). Being a glue language means having APIs to a lot of external toolchains written in not only C/C++ but many other compiled languages, APIs and system resources. Conda, virtualenv, etc. are godsend modules for making it all work, or even better, to freeze things once they all work, without resourcing to Docker, VMs or shell scripts. It's meant for application and DevOps people who need to slap together, ie, ML, Numpy, Elasticsearch, AWS APIs and REST endpoints and Get $hit Done.

    It's annoying to see them "glueys" compared to the binary compiled segment where the heavy lifting is done. Python and others exist to latch on and assimilate. Resistance is futile:

    https://pypi.org/project/pyllamacpp/

    https://www.npmjs.com/package/llama-node

    https://packagist.org/packages/kambo/llama-cpp-php

    https://github.com/yoshoku/llama_cpp.rb

  • collider

    Large Model Collider - The Platform for serving LLM models

  • Project mention: Show HN: Collider – the platform for local LLM debug and inference at warp speed | news.ycombinator.com | 2023-11-30
  • redis-llm

    redis-llm integrates LLM with Redis. It can help LLM to access private data, and remember long chat history.

  • Project mention: Show HN: Redis-LLM – Redis module integrates LLM with Redis | news.ycombinator.com | 2023-07-10
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

C++ llm related posts

Index

What are some of the best open-source llm projects in C++? This list will help you:

Project Stars
1 llama.cpp 56,891
2 LocalAI 19,593
3 PowerInfer 6,942
4 koboldcpp 3,749
5 nitro 1,572
6 rwkv.cpp 1,097
7 distributed-llama 734
8 awadb 158
9 llama_cpp.rb 129
10 collider 117
11 redis-llm 44

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com