ChatGLM, an open-source, self-hosted dialogue language model and alternative to ChatGPT created by Tsinghua University, can be run with as little as 6GB of GPU memory.

This page summarizes the projects mentioned and recommended in the original post on /r/selfhosted

WorkOS - The modern identity platform for B2B SaaS
The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
workos.com
featured
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
  • ChatGLM-6B

    ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型

  • The software itself is licenced under Apache License 2.0, you can always use the software to train your own model if you want to "harm the public interest of society, or infringe upon the rights and interests of human beings".

  • basaran

    Discontinued Basaran is an open-source alternative to the OpenAI text completion API. It provides a compatible streaming API for your Hugging Face Transformers-based text generation models.

  • The ChatGLM image was built using this Dockerfile, basically it's just a "bundled" version of Basaran. The complete usage guide is available here (though not specific to ChatGLM).

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • stable-diffusion-webui

    Stable Diffusion web UI

  • For AI artwork there is Stable Diffusion.

  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

  • This one has an interface similar to A1111.

  • alpaca.cpp

    Discontinued Locally run an Instruction-Tuned Chat-Style LLM

  • This one runs entirely on a CPU. It's a fork of this repo and uses the newly release Alpaca LORA for the LLaMa model.

  • llama.cpp

    LLM inference in C/C++

  • This one runs entirely on a CPU. It's a fork of this repo and uses the newly release Alpaca LORA for the LLaMa model.

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • I made an OpenAI-compatible streaming API (and playground) for your 🤗 Transformers-based text generation models

    2 projects | /r/LanguageTechnology | 6 Mar 2023
  • 1-Jun-2023

    2 projects | /r/dailyainews | 2 Jun 2023
  • Basaran is an open-source alternative to the OpenAI text completion API

    1 project | news.ycombinator.com | 31 May 2023
  • Ask HN: What's the best self hosted/local alternative to GPT-4?

    12 projects | news.ycombinator.com | 31 May 2023
  • Are all the finetunes stupid?

    5 projects | /r/LocalLLaMA | 22 Apr 2023