How do I utilize these quantized models being uploaded?

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • llama.cpp

    LLM inference in C/C++

  • The actual usage is different. For on-CPU inference with ggmlv3.bin models you use llama.cpp or one of its derivatives. GPU-oriented quantization such as GPTQ are also available, but, as I don't have a compatible GPU, I ignore them.

  • koboldcpp

    A simple one-file way to run various GGML and GGUF models with KoboldAI's UI

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

  • ctransformers

    Python bindings for the Transformer models implemented in C/C++ using GGML library.

  • You can also use ctransformers with the ggml models if you want to use python rather than c++.

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Refact LLM: New 1.6B code model reaches 32% HumanEval and is SOTA for the size

    4 projects | news.ycombinator.com | 4 Sep 2023
  • How Open is Generative AI? Part 2

    8 projects | dev.to | 19 Dec 2023
  • Struggling with Local LLMs

    2 projects | /r/artificial | 4 Jul 2023
  • Local LLMs GPUs

    2 projects | /r/LocalLLaMA | 4 Jul 2023
  • Best commercially viable method to ask questions against a set of 30~ PDFs?

    3 projects | /r/LocalLLaMA | 28 Jun 2023