GPTQ-for-LLaMa VS Local-LLM-Comparison-Colab-UI

Compare GPTQ-for-LLaMa vs Local-LLM-Comparison-Colab-UI and see what are their differences.

GPTQ-for-LLaMa

4 bits quantization of LLaMa using GPTQ (by oobabooga)

Local-LLM-Comparison-Colab-UI

Compare the performance of different LLM that can be deployed locally on consumer hardware. Run yourself with Colab WebUI. (by Troyanovsky)
Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
GPTQ-for-LLaMa Local-LLM-Comparison-Colab-UI
19 20
129 868
- -
7.7 9.1
11 months ago 3 days ago
Python Jupyter Notebook
- -
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

GPTQ-for-LLaMa

Posts with mentions or reviews of GPTQ-for-LLaMa. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-06-11.

Local-LLM-Comparison-Colab-UI

Posts with mentions or reviews of Local-LLM-Comparison-Colab-UI. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-07-06.
  • Mistral 7B OpenOrca outclasses Llama 2 13B variants
    1 project | news.ycombinator.com | 21 Oct 2023
  • GPT-4 API general availability
    15 projects | news.ycombinator.com | 6 Jul 2023
    In terms of speed, we're talking about 140t/s for 7B models, and 40t/s for 33B models on a 3090/4090 now.[1] (1 token ~= 0.75 word) It's quite zippy. llama.cpp performs close on Nvidia GPUs now (but they don't have a handy chart) and you can get decent performance on 13B models on M1/M2 Macs.

    You can take a look at a list of evals here: https://llm-tracker.info/books/evals/page/list-of-evals - for general usage, I think home-rolled evals like llm-jeopardy [2] and local-llm-comparison [3] by hobbyists are more useful than most of the benchmark rankings.

    That being said, personally I mostly use GPT-4 for code assistance to that's what I'm most interested in, and the latest code assistants are scoring quite well: https://github.com/abacaj/code-eval - a recent replit-3b fine tune the human-eval results for open models (as a point of reference, GPT-3.5 gets 60.4 on pass@1 and 68.9 on pass@10 [4]) - I've only just started playing around with it since replit model tooling is not as good as llamas (doc here: https://llm-tracker.info/books/howto-guides/page/replit-mode...).

    I'm interested in potentially applying reflexion or some of the other techniques that have been tried to even further increase coding abilities. (InterCode in particular has caught my eye https://intercode-benchmark.github.io/)

    [1] https://github.com/turboderp/exllama#results-so-far

    [2] https://github.com/aigoopy/llm-jeopardy

    [3] https://github.com/Troyanovsky/Local-LLM-comparison/tree/mai...

    [4] https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder

  • Best 7B model
    1 project | /r/oobaboogazz | 29 Jun 2023
    The best 7B I tried is WizardLM. It's my go-to model.
  • UltraLM-13B reaches top of AlpacaEval leaderboard
    3 projects | /r/LocalLLaMA | 28 Jun 2023
    If you want to try it out, you can use Google Colab here with Oobabooga Text Generation UI: Link (Remember to check the instruction template and generation parameters)
  • wizardLM-7B.q4_2
    1 project | /r/LocalLLaMA | 18 Jun 2023
    I'm really impressed by wizardLM-7B.q4_2 (GPT4all) running on my 8gb M2 Mac Air. Fast response, fewer hallucinations than other 7B models I've tried. GPT4All's beta document collection and query function is respectable--going to test it more tomorrow. FWIW wizardLM-7B.q4_2 was ranked very high here https://github.com/Troyanovsky/Local-LLM-comparison.
  • Help me discover new LLMs for school project
    4 projects | /r/LocalLLaMA | 18 Jun 2023
    I made a series of Colab notebooks for different models: https://github.com/Troyanovsky/Local-LLM-comparison
  • Nous Hermes 13b is very good.
    1 project | /r/LocalLLaMA | 11 Jun 2023
    I found it performing very well too in my testing (Repo). It's my second favorite model after WizardLM-13B.
  • How to train 7B models with small documents?
    2 projects | /r/LocalLLaMA | 9 Jun 2023
  • What are your favorite LLMs?
    4 projects | /r/LocalLLaMA | 8 Jun 2023
    My entire list at: Local LLM Comparison Repo
  • Announcing Nous-Hermes-13b (info link in thread)
    3 projects | /r/LocalLLaMA | 3 Jun 2023
    I just tried HyperMantis and updated the results in the repo. It performs not bad but worse than Nous-Hermes-13B.

What are some alternatives?

When comparing GPTQ-for-LLaMa and Local-LLM-Comparison-Colab-UI you can also consider the following projects:

exllama - A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.

langflow - ⛓️ Langflow is a dynamic graph where each node is an executable unit. Its modular and interactive design fosters rapid experimentation and prototyping, pushing hard on the limits of creativity.

koboldcpp - A simple one-file way to run various GGML and GGUF models with KoboldAI's UI

private-gpt - Interact with your documents using the power of GPT, 100% privately, no data leaks

simple-proxy-for-tavern

GPTQ-for-LLaMa - 4 bits quantization of LLaMA using GPTQ

one-click-installers - Simplified installers for oobabooga/text-generation-webui.

alpaca_eval - An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.

can-ai-code - Self-evaluating interview for AI coders