mpt-lora-patch
Patch for MPT-7B which allows using and training a LoRA (by iwalton3)
LoRA-MPT
A repo for finetuning MPT using LoRA. It is currently configured to work with the Alpaca dataset from Stanford but can easily be adapted to use another. (by mikeybellissimo)
mpt-lora-patch | LoRA-MPT | |
---|---|---|
4 | 1 | |
58 | 18 | |
- | - | |
5.1 | 5.9 | |
about 1 year ago | 12 months ago | |
Python | Jupyter Notebook | |
- | Apache License 2.0 |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
mpt-lora-patch
Posts with mentions or reviews of mpt-lora-patch.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-05-29.
- Anyone here finetune either MPT-7B or Falcon-7B?
-
How big of a jump is 13B Vicuna Uncensored vs 30B Vicuna Uncensored?
To merge a LoRA into an existing model, use this script:
-
Training a LoRA with MPT Models
Yeah it's the link on this post https://github.com/iwalton3/mpt-lora-patch
LoRA-MPT
Posts with mentions or reviews of LoRA-MPT.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-05-29.
-
Anyone here finetune either MPT-7B or Falcon-7B?
Update: I successfully trained an alpaca-style model based on mpt storywrite using this repo: https://github.com/mikeybellissimo/LoRA-MPT. Will share more details tommorow.
What are some alternatives?
When comparing mpt-lora-patch and LoRA-MPT you can also consider the following projects:
lora-instruct - Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
GPTQ-for-LLaMa - 4 bits quantization of LLaMA using GPTQ
AGiXT - AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.