Up your coding game and discover issues early. SonarLint is a free plugin that helps you find & fix bugs and security issues from the moment you start writing code. Install from your favorite IDE marketplace today. Learn more →
Similar projects and alternatives to alpaca-lora
A gradio web UI for running Large Language Models like GPT-J 6B, OPT, GALACTICA, LLaMA, and Pygmalion.
Port of Facebook's LLaMA model in C/C++
Access the most powerful time series database as a service. Ingest, store, & analyze all types of time series data in a fully-managed, purpose-built database. Keep data forever with low-cost storage and superior data compression.
Code and documentation to train Stanford's Alpaca models, and generate the data.
Locally run an Instruction-Tuned Chat-Style LLM
Inference code for LLaMA models
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
The simplest way to run LLaMA on your local machine
Clean code begins in your IDE with SonarLint. Up your coding game and discover issues early. SonarLint is a free plugin that helps you find & fix bugs and security issues from the moment you start writing code. Install from your favorite IDE marketplace today.
🌸 Run 100B+ language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models" (by microsoft)
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Alpaca-LoRA as Chatbot service
Plug and play modules to optimize the performances of your AI systems 🚀
Train transformer language models with reinforcement learning.
Repository for training a LoRA for the LLaMA model on HuggingFace with 8-bit quantization. Research only.
Model parallel transformers in JAX and Haiku
ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
Using Low-rank adaptation to quickly fine-tune diffusion models.
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
alpaca-lora reviews and mentions
LoRA: Low-Rank Adaptation of Large Language Models
6 projects | news.ycombinator.com | 24 Mar 2023
For those wondering why this is interesting: This technique is being used to reproduce (though unclear exactly with what fidelity) the Alpaca results from Stanford with a few hours of training on consumer-grade hardware.
I believe that there will be a cottage industry of providing application-specific fine-tuned models like this, that can run in e.g. AWS very inexpensively. The barrier today seems to be that the base model (here, LLaMa) is encumbered and can't be used commercially. Someone will soon, I'm confident, release e.g. an MIT-licensed equivalent and we'll all be off to the races.
Show HN: Web demo of 13B Alpaca-LLaMA trained on improved Stanford dataset
3 projects | news.ycombinator.com | 23 Mar 2023
- Fine-tuning dataset was based on https://github.com/tloen/alpaca-lora/raw/81eb72f707b0505a03b... with minor improvements; I'm going to publish my version soon
Alpaca-LoRA: Low-Rank LLaMA Instruct-Tuning
2 projects | reddit.com/r/LocalLLaMA | 23 Mar 2023
I wrote a Discord bot to host your own ChatGPT-style chatbot with ALPACA finetuned LLaMA weights on consumer GPUs. 13b fits in a 3080, 30b fits in a 3090. No censorship, ask anything you want!
2 projects | reddit.com/r/ChatGPT | 22 Mar 2023
Show HN: ChatLLaMA – A ChatGPT style chatbot for Facebook's LLaMA
10 projects | news.ycombinator.com | 22 Mar 2023
did you use the cleaned and improved alpaca dataset from https://github.com/tloen/alpaca-lora/issues/28 ?
[P] fastLLaMa, A python wrapper to run llama.cpp
4 projects | reddit.com/r/MachineLearning | 21 Mar 2023
Amazing work! Any plans to get this working with alpaca-lora? Would love to see your fast implementation have the improved outputs from Alpaca.
Show HN: Finetune LLaMA-7B on commodity GPUs using your own text
16 projects | news.ycombinator.com | 21 Mar 2023
I've been playing around with https://github.com/zphang/minimal-llama/ and https://github.com/tloen/alpaca-lora/blob/main/finetune.py, and wanted to create a simple UI where you can just paste text, tweak the parameters, and finetune the model quickly using a modern GPU.
To prepare the data, simply separate your text with two blank lines.
There's an inference tab, so you can test how the tuned model behaves.
This is my first foray into the world of LLM finetuning, Python, Torch, Transformers, LoRA, PEFT, and Gradio.
How to do Llama 30B 4bit finetuning?
4 projects | reddit.com/r/LocalLLaMA | 21 Mar 2023
alpaca-lora applied this successfully to fine-tuning LLaMa, and then exported / combined with the original model, later quantizing back to 4-bit so that it could be loaded by alpaca.cpp.4 projects | reddit.com/r/LocalLLaMA | 21 Mar 2023
As for the Lora the 7b trained on a 4090 16gb in 5hrs https://github.com/tloen/alpaca-lora, and almost ran out of memory. I’ve read the 30b has been trained on a single a100 with 40gb
It can be done! (Devs attention required)
5 projects | reddit.com/r/PygmalionAI | 21 Mar 2023
A note from our sponsor - SonarLint
www.sonarlint.org | 25 Mar 2023
tloen/alpaca-lora is an open source project licensed under Apache License 2.0 which is an OSI approved license.