text-generation-webui VS stanford_alpaca

Compare text-generation-webui vs stanford_alpaca and see what are their differences.

text-generation-webui

A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models. (by oobabooga)

stanford_alpaca

Code and documentation to train Stanford's Alpaca models, and generate the data. (by tatsu-lab)
Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
text-generation-webui stanford_alpaca
876 108
35,862 28,761
- 1.3%
9.9 2.0
4 days ago about 1 month ago
Python Python
GNU Affero General Public License v3.0 Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

text-generation-webui

Posts with mentions or reviews of text-generation-webui. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-01.
  • Ask HN: What is the current (Apr. 2024) gold standard of running an LLM locally?
    11 projects | news.ycombinator.com | 1 Apr 2024
    Some of the tools offer a path to doing tool use (fetching URLs and doing things with them) or RAG (searching your documents). I think Oobabooga https://github.com/oobabooga/text-generation-webui offers the latter through plugins.

    Our tool, https://github.com/transformerlab/transformerlab-app also supports the latter (document search) using local llms.

  • Ask HN: How to get started with local language models?
    6 projects | news.ycombinator.com | 17 Mar 2024
    You can use webui https://github.com/oobabooga/text-generation-webui

    Once you get a version up and running I make a copy before I update it as several times updates have broken my working version and caused headaches.

    a decent explanation of parameters outside of reading archive papers: https://github.com/oobabooga/text-generation-webui/wiki/03-%...

    a news ai website:

  • text-generation-webui VS LibreChat - a user suggested alternative
    2 projects | 29 Feb 2024
  • Show HN: I made an app to use local AI as daily driver
    31 projects | news.ycombinator.com | 27 Feb 2024
  • Ask HN: People who switched from GPT to their own models. How was it?
    3 projects | news.ycombinator.com | 26 Feb 2024
    The other answers are recommending paths which give you #1. less control and #2. projects with smaller eco-systems.

    If you want a truly general purpose front-end for LLMs, the only good solution right now is oobabooga: https://github.com/oobabooga/text-generation-webui

    All other alternatives have only small fractions of the features that oobabooga supports. All other alternatives only support a fraction of the LLM backends that oobabooga supports, etc.

  • AI Girlfriend Is a Data-Harvesting Horror Show
    1 project | news.ycombinator.com | 14 Feb 2024
    The example waifu in text-generation-webui is good enough for me.

    https://github.com/oobabooga/text-generation-webui/blob/main...

  • Nvidia's Chat with RTX is a promising AI chatbot that runs locally on your PC
    7 projects | news.ycombinator.com | 13 Feb 2024
    > Downloading text-generation-webui takes a minute, let's you use any model and get going.

    What you're missing here is you're already in this area deep enough to know what ooogoababagababa text-generation-webui is. Let's back out to the "average Windows desktop user" level. Assuming they even know how to find it:

    1) Go to https://github.com/oobabooga/text-generation-webui?tab=readm...

    2) See a bunch of instructions opening a terminal window and running random batch/powershell scripts. Powershell, etc will likely prompt you with a scary warning. Then you start wondering who ooobabagagagaba is...

    3) Assuming you get this far (many users won't even get to step 1) you're greeted with a web interface[0] FILLED to the brim with technical jargon and extremely overwhelming options just to get a model loaded, which is another mind warp because you get to try to select between a bunch of random models with no clear meaning and non-sensical/joke sounding names from someone called "TheBloke". Ok...

    Let's say you somehow braved this gauntlet and get this far now you get to chat with it. Ok, what about my local documents? text-generation-webui itself has nothing for that. Repeat this process over the 10 random open source projects from a bunch of names you've never heard of in an attempt to accomplish that.

    This is "I saw this thing from Nvidia explode all over media, twitter, youtube, etc. I downloaded it from Nvidia, double-clicked, pointed it at a folder with documents, and it works".

    That's the difference and it's very significant.

    [0] - https://raw.githubusercontent.com/oobabooga/screenshots/main...

  • Ask HN: What are your top 3 coolest software engineering tools?
    1 project | news.ycombinator.com | 6 Feb 2024
    Maybe a copout answer, but setting up a local LLM on my development machine has been invaluable. I use Deep Seek Coder 6.7 [0] and Oobabooga's UI [1]. It helps me solve simple problems and find bugs, while still leaving the larger architecture decisions to me.

    [0] https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instr...

    [1] https://github.com/oobabooga/text-generation-webui

  • Meta AI releases Code Llama 70B
    6 projects | news.ycombinator.com | 29 Jan 2024
    You can download it and run it with [this](https://github.com/oobabooga/text-generation-webui). There's an API mode that you could leverage from your VS Code extension.
  • Ollama Python and JavaScript Libraries
    17 projects | news.ycombinator.com | 24 Jan 2024
    Same question here. Ollama is fantastic as it makes it very easy to run models locally, But if you already have a lot of code that processes OpenAI API responses (with retry, streaming, async, caching etc), it would be nice to be able to simply switch the API client to Ollama, without having to have a whole other branch of code that handles Alama API responses. One way to do an easy switch is using the litellm library as a go-between but it’s not ideal (and I also recently found issues with their chat formatting for mistral models).

    For an OpenAI compatible API my current favorite method is to spin up models using oobabooga TGW. Your OpenAI API code then works seamlessly by simply switching out the api_base to the ooba endpoint. Regarding chat formatting, even ooba’s Mistral formatting has issues[1] so I am doing my own in Langroid using HuggingFace tokenizer.apply_chat_template [2]

    [1] https://github.com/oobabooga/text-generation-webui/issues/53...

    [2] https://github.com/langroid/langroid/blob/main/langroid/lang...

    Related question - I assume ollama auto detects and applies the right chat formatting template for a model?

stanford_alpaca

Posts with mentions or reviews of stanford_alpaca. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-12-19.
  • How Open is Generative AI? Part 2
    8 projects | dev.to | 19 Dec 2023
    Alpaca is an instruction-oriented LLM derived from LLaMA, enhanced by Stanford researchers with a dataset of 52,000 examples of following instructions, sourced from OpenAI’s InstructGPT through the self-instruct method. The extensive self-instruct dataset, details of data generation, and the model refinement code were publicly disclosed. This model complies with the licensing requirements of its base model. Due to the utilization of InstructGPT for data generation, it also adheres to OpenAI’s usage terms, which prohibit the creation of models competing with OpenAI. This illustrates how dataset restrictions can indirectly affect the resulting fine-tuned model.
  • Ask HN: AI/ML papers to catch up with current state of AI?
    3 projects | news.ycombinator.com | 15 Dec 2023
  • OpenAI board in discussions with Sam Altman to return as CEO
    1 project | news.ycombinator.com | 19 Nov 2023
  • Are there any AI like ChatGPT without content restrictions?
    1 project | /r/OpenAI | 3 Oct 2023
  • Fine-tuning LLMs with LoRA: A Gentle Introduction
    3 projects | dev.to | 22 Aug 2023
    In this article, we're going to experiment with LoRA and fine-tune Llama Alpaca using commercial hardware.
  • Creating a new Finetuned model
    3 projects | /r/LocalLLaMA | 11 Jul 2023
    Most papers I did read showed at least a thousand, even 10000 at several cases, so I assumed that to be the trend in the case of Low rank adapter(PEFT) training.(source: [2305.14314] QLoRA: Efficient Finetuning of Quantized LLMs (arxiv.org) , Stanford CRFM (Alpaca) and the minimum being openchat/openchat · Hugging Face ; There are a lot more examples)
  • Shock tick up for wage growth to 7.3% in blow for Bank of England
    1 project | /r/unitedkingdom | 11 Jul 2023
    I'm not talking about OpenAI ChatGPT I'm talking about things ALPACA, and where did they train these models? Off the existing models for a fraction of a fraction of a fraction of the cost: https://crfm.stanford.edu/2023/03/13/alpaca.html
  • Bye bye Bing
    5 projects | /r/ChatGPT | 30 Jun 2023
  • The idea maze for AI startups (2015)
    2 projects | news.ycombinator.com | 28 Jun 2023
    I think there's a new approach for “How do you get the data?” that wasn't available when this article was written in 2015. The new text and image generative models can now be used to synthesize training datasets.

    I was working on an typing autocorrect project and needed a corpus of "text messages". Most of the traditional NLP corpuses like those available through NLTK [0] aren't suitable. But it was easy to script ChatGPT to generate thousands of believable text messages by throwing random topics at it.

    Similarly, you can synthesize a training dataset by giving GPT the outputs/labels and asking it to generate a variety of inputs. For sentiment analysis... "Give me 1000 negative movie reviews" and "Now give me 1000 positive movie reviews".

    The Alpaca folks used GPT-3 to generate high-quality instruction-following datasets [1] based on a small set of human samples.

    Etc.

    [0] https://www.nltk.org/nltk_data/

    [1] https://crfm.stanford.edu/2023/03/13/alpaca.html

  • Repos and tutorials for a full finetune (not LoRA)
    1 project | /r/LocalLLaMA | 2 Jun 2023
    AFAIK, the original alpaca repo was a full finetune. https://github.com/tatsu-lab/stanford_alpaca

What are some alternatives?

When comparing text-generation-webui and stanford_alpaca you can also consider the following projects:

KoboldAI

alpaca-lora - Instruct-tune LLaMA on consumer hardware

llama.cpp - LLM inference in C/C++

ChatGLM-6B - ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型

gpt4all - gpt4all: run open-source LLMs anywhere

Open-Assistant - OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.

TavernAI - Atmospheric adventure chat for AI language models (KoboldAI, NovelAI, Pygmalion, OpenAI chatgpt, gpt-4)

KoboldAI-Client

GPTQ-for-LLaMa - 4 bits quantization of LLaMA using GPTQ

ollama - Get up and running with Llama 2, Mistral, Gemma, and other large language models.

Alpaca-Turbo - Web UI to run alpaca model locally