llm

Top 23 llm Open-Source Projects

  • llama.cpp

    LLM inference in C/C++

  • Project mention: Better and Faster Large Language Models via Multi-Token Prediction | news.ycombinator.com | 2024-05-01

    For anyone interested in exploring this, llama.cpp has an example implementation here:

    https://github.com/ggerganov/llama.cpp/tree/master/examples/...

  • MetaGPT

    🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming

  • Project mention: Can AI replace a co-founder? | news.ycombinator.com | 2024-01-07

    https://github.com/geekan/MetaGPT :

    > MetaGPT takes a one line requirement as input and outputs user stories / competitive analysis / requirements / data structures / APIs / documents, etc.

    https://news.ycombinator.com/item?id=29141796 ; "Co-Founder Equity Calculator"

    "Ask HN: What are your go to SaaS products for startups/MVPs?" (2020) https://news.ycombinator.com/item?id=23535828 ; FounderKit, StackShare

    > USA Small Business Administration: "10 steps to start your business." https://www.sba.gov/starting-business/how-start-business/10-...

    >> "Startup Incorporation Checklist: How to bootstrap a Delaware C-corp (or S-corp) with employee(s) in California" https://github.com/leonar15/startup-checklist

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • llama_index

    LlamaIndex is a data framework for your LLM applications

  • Project mention: LlamaIndex: A data framework for your LLM applications | news.ycombinator.com | 2024-04-07
  • llm-course

    Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.

  • Project mention: Ask HN: People who switched from GPT to their own models. How was it? | news.ycombinator.com | 2024-02-26

    This is a very nice resource: https://github.com/mlabonne/llm-course

  • Milvus

    A cloud-native vector database, storage for next generation AI applications

  • Project mention: Ask HN: Who is hiring? (April 2024) | news.ycombinator.com | 2024-04-01

    Zilliz (zilliz.com) | Hybrid/ONSITE (SF, NYC) | Full-time

    I am part of the hiring team for DevRel

    NYC - https://boards.greenhouse.io/zilliz/jobs/4307910005

    SF - https://boards.greenhouse.io/zilliz/jobs/4317590005

    Zilliz is the company behind Milvus (https://github.com/milvus-io/milvus), the most starred vector database on GitHub. Milvus is a distributed vector database that shines in 1B+ vector use cases. Examples include autonomous driving, e-commerce, and drug discovery. (and, of course, RAG)

    We are also hiring for other roles that I am not personally involved in the hiring process for such as product managers, software engineers, and recruiters.

  • dify

    Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.

  • Project mention: FLaNK AI Weekly for 29 April 2024 | dev.to | 2024-04-29
  • chatgpt-on-wechat

    基于大模型搭建的聊天机器人,同时支持 企业微信、微信 公众号、飞书、钉钉 等接入,可选择GPT3.5/GPT4.0/Claude/文心一言/讯飞星火/通义千问/Gemini/GLM-4/Claude/Kimi/LinkAI,能处理文本、语音和图片,访问操作系统和互联网,支持基于自有知识库进行定制企业智能客服。

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
  • Flowise

    Drag & drop UI to build your customized LLM flow

  • Project mention: FLaNK Stack Weekly 12 February 2024 | dev.to | 2024-02-12
  • MindsDB

    The platform for customizing AI from enterprise data

  • Project mention: What’s the Difference Between Fine-tuning, Retraining, and RAG? | dev.to | 2024-04-08

    Check us out on GitHub.

  • LLaMA-Factory

    Unify Efficient Fine-Tuning of 100+ LLMs

  • Project mention: Show HN: GPU Prices on eBay | news.ycombinator.com | 2024-02-23

    Depends what model you want to train, and how well you want your computer to keep working while you're doing it.

    If you're interested in large language models there's a table of vram requirements for fine-tuning at [1] which says you could do the most basic type of fine-tuning on a 7B parameter model with 8GB VRAM.

    You'll find that training takes quite a long time, and as a lot of the GPU power is going on training, your computer's responsiveness will suffer - even basic things like scrolling in your web browser or changing tabs uses the GPU, after all.

    Spend a bit more and you'll probably have a better time.

    [1] https://github.com/hiyouga/LLaMA-Factory?tab=readme-ov-file#...

  • LocalAI

    :robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.

  • Project mention: Drop-In Replacement for ChatGPT API | news.ycombinator.com | 2024-01-24
  • vllm

    A high-throughput and memory-efficient inference and serving engine for LLMs

  • Project mention: AI leaderboards are no longer useful. It's time to switch to Pareto curves | news.ycombinator.com | 2024-04-30

    I guess the root cause of my claim is that OpenAI won't tell us whether or not GPT-3.5 is an MoE model, and I assumed it wasn't. Since GPT-3.5 is clearly nondeterministic at temp=0, I believed the nondeterminism was due to FPU stuff, and this effect was amplified with GPT-4's MoE. But if GPT-3.5 is also MoE then that's just wrong.

    What makes this especially tricky is that small models are truly 100% deterministic at temp=0 because the relative likelihoods are too coarse for FPU issues to be a factor. I had thought 3.5 was big enough that some of its token probabilities were too fine-grained for the FPU. But that's probably wrong.

    On the other hand, it's not just GPT, there are currently floating-point difficulties in vllm which significantly affect the determinism of any model run on it: https://github.com/vllm-project/vllm/issues/966 Note that a suggested fix is upcasting to float32. So it's possible that GPT-3.5 is using an especially low-precision float and introducing nondeterminism by saving money on compute costs.

    Sadly I do not have the money[1] to actually run a test to falsify any of this. It seems like this would be a good little research project.

    [1] Or the time, or the motivation :) But this stuff is expensive.

  • unilm

    Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities

  • Project mention: The Era of 1-Bit LLMs: Training_Tips, Code And_FAQ [pdf] | news.ycombinator.com | 2024-03-21
  • semantic-kernel

    Integrate cutting-edge LLM technology quickly and easily into your apps

  • Project mention: #SemanticKernel – 📎Chat Service demo running Phi-2 LLM locally with #LMStudio | dev.to | 2024-02-08

    There is an amazing sample on how to create your own LLM Service class to be used in Semantic Kernel. You can view the Sample here: https://github.com/microsoft/semantic-kernel/blob/3451a4ebbc9db0d049f48804c12791c681a326cb/dotnet/samples/KernelSyntaxExamples/Example16_CustomLLM.cs

  • Chinese-LLaMA-Alpaca

    中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)

  • Project mention: Chinese-Alpaca-Plus-13B-GPTQ | /r/LocalLLaMA | 2023-05-30

    I'd like to share with you today the Chinese-Alpaca-Plus-13B-GPTQ model, which is the GPTQ format quantised 4bit models of Yiming Cui's Chinese-LLaMA-Alpaca 13B for GPU reference.

  • mlc-llm

    Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.

  • Project mention: FLaNK 04 March 2024 | dev.to | 2024-03-04
  • open-webui

    User-friendly WebUI for LLMs (Formerly Ollama WebUI)

  • Project mention: open-webui VS LibreChat - a user suggested alternative | libhunt.com/r/open-webui | 2024-02-29
  • ChatGLM2-6B

    ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型

  • Project mention: Are We Overlooking China's Progress in AI? | /r/singularity | 2023-06-26
  • LLMs-from-scratch

    Implementing a ChatGPT-like LLM from scratch, step by step

  • Project mention: Insights from Finetuning LLMs for Classification Tasks | news.ycombinator.com | 2024-04-28
  • peft

    🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

  • Project mention: LoftQ: LoRA-fine-tuning-aware Quantization | news.ycombinator.com | 2023-12-19
  • dalai

    The simplest way to run LLaMA on your local machine

  • Project mention: Ask HN: What are the capabilities of consumer grade hardware to work with LLMs? | news.ycombinator.com | 2023-08-03

    I agree, I've definitely seen way more information about running image synthesis models like Stable Diffusion locally than I have LLMs. It's counterintuitive to me that Stable Diffusion takes less RAM than an LLM, especially considering it still needs the word vectors. Goes to show I know nothing.

    I guess it comes down to the requirement of a very high end (or multiple) GPU that makes it impractical for most vs just running it in Colab or something.

    Tho there are some efforts:

    https://github.com/cocktailpeanut/dalai

  • FastGPT

    FastGPT is a knowledge-based platform built on the LLMs, offers a comprehensive suite of out-of-the-box capabilities such as data processing, RAG retrieval, and visual AI workflow orchestration, letting you easily develop and deploy complex question-answering systems without the need for extensive setup or configuration.

  • Project mention: FLaNK Stack Weekly 12 February 2024 | dev.to | 2024-02-12
  • anything-llm

    The all-in-one Desktop & Docker AI application with full RAG and AI Agent capabilities.

  • Project mention: AnythingLLM: Chat with your documents using any LLM | news.ycombinator.com | 2024-04-19
  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

llm related posts

Index

What are some of the best open-source llm projects? This list will help you:

Project Stars
1 llama.cpp 56,891
2 MetaGPT 39,302
3 llama_index 31,184
4 llm-course 28,809
5 Milvus 26,857
6 dify 25,645
7 chatgpt-on-wechat 24,945
8 Flowise 24,074
9 MindsDB 21,312
10 LLaMA-Factory 20,248
11 LocalAI 19,862
12 vllm 18,571
13 unilm 18,358
14 semantic-kernel 18,234
15 Chinese-LLaMA-Alpaca 17,348
16 mlc-llm 16,955
17 open-webui 16,677
18 ChatGLM2-6B 15,495
19 LLMs-from-scratch 14,142
20 peft 13,877
21 dalai 13,051
22 FastGPT 12,961
23 anything-llm 12,420

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com