stanford_alpaca VS LoRA

Compare stanford_alpaca vs LoRA and see what are their differences.

stanford_alpaca

Code and documentation to train Stanford's Alpaca models, and generate the data. (by tatsu-lab)

LoRA

Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models" (by microsoft)
Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
stanford_alpaca LoRA
108 34
28,723 8,890
1.2% 7.0%
2.0 5.4
about 1 month ago about 1 month ago
Python Python
Apache License 2.0 MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

stanford_alpaca

Posts with mentions or reviews of stanford_alpaca. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-12-19.
  • How Open is Generative AI? Part 2
    8 projects | dev.to | 19 Dec 2023
    Alpaca is an instruction-oriented LLM derived from LLaMA, enhanced by Stanford researchers with a dataset of 52,000 examples of following instructions, sourced from OpenAI’s InstructGPT through the self-instruct method. The extensive self-instruct dataset, details of data generation, and the model refinement code were publicly disclosed. This model complies with the licensing requirements of its base model. Due to the utilization of InstructGPT for data generation, it also adheres to OpenAI’s usage terms, which prohibit the creation of models competing with OpenAI. This illustrates how dataset restrictions can indirectly affect the resulting fine-tuned model.
  • Ask HN: AI/ML papers to catch up with current state of AI?
    3 projects | news.ycombinator.com | 15 Dec 2023
  • Fine-tuning LLMs with LoRA: A Gentle Introduction
    3 projects | dev.to | 22 Aug 2023
    In this article, we're going to experiment with LoRA and fine-tune Llama Alpaca using commercial hardware.
  • Creating a new Finetuned model
    3 projects | /r/LocalLLaMA | 11 Jul 2023
    Most papers I did read showed at least a thousand, even 10000 at several cases, so I assumed that to be the trend in the case of Low rank adapter(PEFT) training.(source: [2305.14314] QLoRA: Efficient Finetuning of Quantized LLMs (arxiv.org) , Stanford CRFM (Alpaca) and the minimum being openchat/openchat · Hugging Face ; There are a lot more examples)
  • Bye bye Bing
    5 projects | /r/ChatGPT | 30 Jun 2023
  • The idea maze for AI startups (2015)
    2 projects | news.ycombinator.com | 28 Jun 2023
    I think there's a new approach for “How do you get the data?” that wasn't available when this article was written in 2015. The new text and image generative models can now be used to synthesize training datasets.

    I was working on an typing autocorrect project and needed a corpus of "text messages". Most of the traditional NLP corpuses like those available through NLTK [0] aren't suitable. But it was easy to script ChatGPT to generate thousands of believable text messages by throwing random topics at it.

    Similarly, you can synthesize a training dataset by giving GPT the outputs/labels and asking it to generate a variety of inputs. For sentiment analysis... "Give me 1000 negative movie reviews" and "Now give me 1000 positive movie reviews".

    The Alpaca folks used GPT-3 to generate high-quality instruction-following datasets [1] based on a small set of human samples.

    Etc.

    [0] https://www.nltk.org/nltk_data/

    [1] https://crfm.stanford.edu/2023/03/13/alpaca.html

  • [D] High-quality, open-source implementations of LLMs
    6 projects | /r/MachineLearning | 22 May 2023
    Alpaca [GitHub]
  • please 0.1.0 released: let GPT-4 remember CLI args
    2 projects | /r/rust | 21 May 2023
    Now if only this could be used offline, eg. with alpaca https://github.com/tatsu-lab/stanford_alpaca
  • Is there a Chatgpt (or other LLMs) powered application in the field of cybersecurity/privacy for end users/b2c?
    2 projects | /r/privacy | 19 May 2023
    If you have a strong enough computer, there is Alpaca and llama.cpp which are both open-source. They also have the best privacy feature of all: to be able to be ran locally offline on your computer. I believe there are more foss LLMs out there too but idr.
  • Does ChatGPT suck at programming for everyone or just for me?
    2 projects | /r/ChatGPT | 15 May 2023
    Are you aware that you can run a pretrained LLM on just 8gb of ram with a single x86 cpu?

LoRA

Posts with mentions or reviews of LoRA. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-12-08.
  • Can a LoRa be used on models other than Stable Diffusion?
    2 projects | /r/StableDiffusion | 8 Dec 2023
    LoRA was initially developed for large language models, https://arxiv.org/abs/2106.09685 (2021). It was later that people discovered that it worked REALLY well for diffusion models.
    2 projects | /r/StableDiffusion | 8 Dec 2023
  • StyleTTS2 – open-source Eleven Labs quality Text To Speech
    10 projects | news.ycombinator.com | 19 Nov 2023
    Curious if we'll see a Civitai-style LoRA[1] marketplace for text-to-speech models.

    1 = https://github.com/microsoft/LoRA

  • Is supervised learning dead for computer vision?
    9 projects | news.ycombinator.com | 28 Oct 2023
    Yes, your understanding is correct. However, instead of adding a head on top of the network, most fine-tuning is currently done with LoRA (https://github.com/microsoft/LoRA). This introduces low-rank matrices between different layers of your models, those are then trained using your training data while the rest of the models' weights are frozen.
  • Run LLMs at home, BitTorrent‑style
    10 projects | news.ycombinator.com | 17 Sep 2023
    Somewhat yes. See "LoRA": https://arxiv.org/abs/2106.09685

    They're not composable in the sense that you can take these adaptation layers and arbitrarily combine them, but training different models while sharing a common base of weights is a solved problem.

  • Open-source Fine-Tuning on Codebase with Refact
    2 projects | dev.to | 5 Sep 2023
    It's possible to fine-tune all parameters (called "full fine-tune"), but recently PEFT methods became popular. PEFT stands for Parameter-Efficient Fine-Tuning. There are several methods available, the most popular so far is LoRA (2106.09685) that can train less than 1% of the original weights. LoRA has one important parameter -- tensor size, called lora_r. It defines how much information LoRA can add to the network. If your codebase is small, the fine-tuning process will see the same data over and over again, many times in a loop. We found that for a smaller codebase small LoRA tensors work best because it won't overfit as much -- the tensors just don't have the capacity to fit the limited training set exactly. As the codebase gets bigger, tensors should become bigger as well. We also unfreeze token embeddings at a certain codebase size. To pick all the parameters automatically, we have developed a heuristic that calculates a score based on the source files it sees. This score is then used to determine the appropriate LoRA size, number of finetuning steps, and other parameters. We have tested this heuristic on several beta test clients, small codebases of several files, and large codebases like the Linux kernel (consisting of about 50,000 useful source files). If the heuristic doesn't work for you for whatever reason, you can set all the parameters yourself.
  • [P5V6P2] Mother and Daughter (by azfumi)
    2 projects | /r/HonzukiNoGekokujou | 12 Jul 2023
    For your first part of the comment, I can simply refer you to technologies like ControlNet, LoRA and prompt embedding: https://github.com/lllyasviel/ControlNet https://github.com/microsoft/LoRA
  • Your weekly machine learning digest
    2 projects | /r/learnmachinelearning | 3 Jul 2023
  • MEDIAPIPE on-device diffusion plugins for conditioned text-to-image generation
    2 projects | /r/AR_MR_XR | 3 Jul 2023
    Today, we announce MediaPipe diffusion plugins, which enable controllable text-to-image generation to be run on-device. Expanding upon our prior work on GPU inference for on-device large generative models, we introduce new low-cost solutions for controllable text-to-image generation that can be plugged into existing diffusion models and their Low-Rank Adaptation (LoRA) variants.
  • Baize v2 [7B/13B]
    2 projects | /r/LocalLLM | 24 May 2023
    Baize is an open-source chat model trained with LoRA. It uses 100k dialogs generated by letting ChatGPT chat with itself. We also use Alpaca's data to improve its performance. We have released 7B, 13B and 30B models. Please refer to the paper for more details.

What are some alternatives?

When comparing stanford_alpaca and LoRA you can also consider the following projects:

LyCORIS - Lora beYond Conventional methods, Other Rank adaptation Implementations for Stable diffusion.

ComfyUI - The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface.

ControlNet - Let us control diffusion models!

peft - 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

alpaca-lora - Instruct-tune LLaMA on consumer hardware

LLaMA-Adapter - [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters

text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

sd-webui-additional-networks

LLaMA-Adapter - Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters [Moved to: https://github.com/OpenGVLab/LLaMA-Adapter]

gpt4all - gpt4all: run open-source LLMs anywhere

ChatGLM-6B - ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型