Loader Types

This page summarizes the projects mentioned and recommended in the original post on /r/oobaboogazz

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

  • Transformers: the biggest and most famous library for running Large Language Models, and possibly one of the oldest. It was created by a company called Hugging Face, which is where we usually download our models from. It supports many models and has many features but it's slow and wastes GPU memory. https://github.com/huggingface/transformers

  • AutoGPTQ

    An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.

  • AutoGPTQ: an attempt at standardizing GPTQ-for-LLaMa and turning it into a library that is easier to install and use, and that supports more models. https://github.com/PanQiWei/AutoGPTQ

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • exllama

    A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.

  • ExLlama: a meticulously optimized library for running GPTQ models. The author is very knowledgeable in low-level GPU programming, and the result is an implementation that is VERY fast and uses much less memory than GPTQ-for-LLaMa or AutoGPTQ. https://github.com/turboderp/exllama

  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

  • ExLlama_HF: a way to use ExLlama as if it was a transformers model. Transformers implements many parameters like top_k, top_p, etc, that this library reuses without any modifications. It was contributed in a recent PR by Larryvrh: https://github.com/oobabooga/text-generation-webui/pull/2777

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • A look at Apple’s new Transformer-powered predictive text model

    4 projects | news.ycombinator.com | 16 Sep 2023
  • AI enthusiasm #6 - Finetune any LLM you want💡

    2 projects | dev.to | 16 Apr 2024
  • FLaNK Stack 05 Feb 2024

    49 projects | dev.to | 5 Feb 2024
  • Can we discuss MLOps, Deployment, Optimizations, and Speed?

    7 projects | /r/LocalLLaMA | 6 Dec 2023
  • How can I learn more about models, trends, news, etc?

    1 project | /r/LocalLLaMA | 29 Oct 2023