peft
simple-llm-finetuner
peft | simple-llm-finetuner | |
---|---|---|
26 | 12 | |
13,877 | 1,977 | |
4.1% | - | |
9.7 | 10.0 | |
4 days ago | 5 months ago | |
Python | Jupyter Notebook | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
peft
- LoftQ: LoRA-fine-tuning-aware Quantization
-
Fine Tuning Mistral 7B on Magic the Gathering Draft
There is not a lot of great content out there making this clear, but basically all that matters for basic fine tuning is how much VRAM you have -- since the 3090 / 4090 have 24GB VRAM they're both pretty decent fine tuning chips. I think you could probably fine-tune a model up to ~13B parameters on one of them with PEFT (https://github.com/huggingface/peft)
-
Whisper prompt tuning
Hi everyone. Recently I've been looking into the PEFT library (https://github.com/huggingface/peft) and I was wondering if it would be possible to do prompt tuning with OpenAI's Whisper model. They have an example notebook for tuning Whisper with LoRA (https://colab.research.google.com/drive/1vhF8yueFqha3Y3CpTHN6q9EVcII9EYzs?usp=sharing) but I'm not sure how to go about changing it to use prompt tuning instead.
-
Code Llama - The Hugging Face Edition
In the coming days, we'll work on sharing scripts to train models, optimizations for on-device inference, even nicer demos (and for more powerful models), and more. Feel free to like our GitHub repos (transformers, peft, accelerate). Enjoy!
- PEFT 0.5 supports fine-tuning GPTQ models
-
Exploding loss when trying to train OpenOrca-Platypus2-13B
image
-
[D] Is there a difference between p-tuning and prefix tuning ?
I discussed part of this here: https://github.com/huggingface/peft/issues/123
-
How does using QLoRAs when running Llama on CPU work?
It seems like the merge_and_unload function in this PEFT script might be what they are referring to: https://github.com/huggingface/peft/blob/main/src/peft/tuners/lora.py
-
How to merge the two weights into a single weight?
To obtain the original llama model, one may refer to this doc. To merge a lora model with a base model, one may refer to PEFT or use the merge script provided by LMFlow.
-
[D] [LoRA + weight merge every N step] for pre-training?
you could use a callback, like show here, https://github.com/huggingface/peft/issues/286 and call code to merge them here.
simple-llm-finetuner
-
Ask HN: Resource to learn how to train and use ML Models
Just the appropriate reddit groups and follow folks on twitter, plus use a search engine.
1. Learn to run a model, checkout llama.cpp Tons of free models on huggingface.com
2. Learn to finetune a model - https://github.com/lxe/simple-llm-finetuner
3. Learn to train one. PyTorch, TensorFlow, HuggingFace libraries, etc.
Good luck.
- How can I train my custom dataset on top of Vicuna?
-
[D] The best way to train an LLM on company data
So as far as set up goes, you just need to: “”” Git clone https://github.com/lxe/simple-llama-finetuner Cd simple-llama-finetuner Pip install -r requirements.txt Python app.py ## if you’re on a remote machine (Paperspace is my go to) then you may need to edit the last line of this script to set ‘share=True’ in the launch args “””
-
Show HN: Document Q&A with GPT: web, .pdf, .docx, etc.
oobabooga's textgen webui has a tab for fine tuning now. You only need a single consumer GPU to fine tune up to 33B parameter models at a rate of about 200 epochs per hour, per GPU.
There are also one-click finetuning projects which run on free Google Colab GPUs like https://github.com/lxe/simple-llama-finetuner
It's easy and not complex at all.
-
How do I fine tune 4 bit or 8 bit models?
for a single 4090, easiest way to get started and simple to use: https://github.com/lxe/simple-llama-finetuner
- Are there publicly available datasets other than Alpaca that we can use to fine-tune LLaMA?
- Show HN: Finetune LLaMA-7B on commodity GPUs using your own text
- [Project] Finetune LLaMA-7B on commodity GPUs (and Colab) using your own text
What are some alternatives?
lora - Using Low-rank adaptation to quickly fine-tune diffusion models.
alpaca-lora - Instruct-tune LLaMA on consumer hardware
LoRA - Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
paper-qa - LLM Chain for answering questions from documents with citations
Made-With-ML - Learn how to design, develop, deploy and iterate on production-grade ML applications.
dalai - The simplest way to run LLaMA on your local machine
minimal-llama
mlc-llm - Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.
OpenChatKit
minLoRA - minLoRA: a minimal PyTorch library that allows you to apply LoRA to any PyTorch model.
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.