Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Top 16 finetuning Open-Source Projects
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
h2o-llmstudio
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://h2oai.github.io/h2o-llmstudio/
-
xTuring
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
-
DB-GPT-Hub
A repository that contains models, datasets, and fine-tuning techniques for DB-GPT, with the purpose of enhancing model performance in Text-to-SQL
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
Awesome-Text2SQL
Curated tutorials and resources for Large Language Models, Text2SQL, Text2DSL、Text2API、Text2Vis and more.
-
finetune-gpt2xl
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
Project mention: Ask HN: Most efficient way to fine-tune an LLM in 2024? | news.ycombinator.com | 2024-04-04Gemma 7b is 2.4x faster than HF + FA2.
Check out https://github.com/unslothai/unsloth for full benchmarks!
I really like the simplicity of this framework, and they hit on a lot of common problems found in other agent-based frameworks. Most intrigued by the RAG improvements.
Seems like Microsoft was frustrated with the pace of movement in this space and the shitty results of agents (which admittedly kept my interest turned away from agents for the last few months). I'm interested again because it makes practical sense, and from looking at the example notebooks, seems fairly easy to integrate into existing applications.
Maybe this is the 'low code' approach that might actually work, and bridge together engineering and non-engineering resources.
This example was what caught my eye: https://github.com/microsoft/FLAML/blob/main/notebook/autoge...
Project mention: Paid dev gig: develop a basic LLM PEFT finetuning utility | /r/LocalLLaMA | 2023-06-02
Project mention: I'm developing an open-source AI tool called xTuring, enabling anyone to construct a Language Model with just 5 lines of code. I'd love to hear your thoughts! | /r/machinelearningnews | 2023-09-07Explore the project on GitHub here.
Project mention: Show HN: Toolkit for LLM Fine-Tuning, Ablating and Testing | news.ycombinator.com | 2024-04-07
If you can find a large body of good, permissively licensed example code, you can finetune an LLM on it!
There was a similar attempt for Godot script trained a few months ago, and its reportedly pretty good:
https://github.com/minosvasilias/godot-dodo
I think more attempts havent been made because base llama is not that great at coding in general, relative to its other strengths, and stuff like Starcoder has flown under the radar.
You can use the train script here https://github.com/kuutsav/llm-toys/blob/main/llm_toys/train.py. The readme contains a sample training command.
Project mention: A full tutorial on turning GPT-2 into a conversational AI | news.ycombinator.com | 2023-08-31Hi, Vatsa here, this is tutorial on turning GPT-2 into a conversational bot, it was a fun project, and I hope you like it it!
github -> https://github.com/VatsaDev/nanoChatGPT
Project mention: Zicklein - a German 🇩🇪 finetuned LlaMA-7b base model (OS) | /r/LocalLLaMA | 2023-05-22
finetuning related posts
-
Ask HN: Most efficient way to fine-tune an LLM in 2024?
-
AMD ROCm Software Blogs
-
Show HN: We got fine-tuning Mistral-7B to not suck
-
Mistral 7B Fine-Tune Optimized
-
Has anyone tried out the ASPEN-Framework for LoRA Fine-Tuning yet and can share their experience?
-
Show HN: 80% faster, 50% less memory, 0% loss of accuracy Llama finetuning
-
80% faster, 50% less memory, 0% loss of accuracy Llama finetuning
-
A note from our sponsor - InfluxDB
www.influxdata.com | 15 May 2024
Index
What are some of the best open-source finetuning projects? This list will help you:
Project | Stars | |
---|---|---|
1 | unsloth | 8,974 |
2 | FLAML | 3,690 |
3 | h2o-llmstudio | 3,614 |
4 | learn2learn | 2,552 |
5 | xTuring | 2,525 |
6 | finetuner | 1,432 |
7 | DB-GPT-Hub | 1,069 |
8 | Awesome-Text2SQL | 1,060 |
9 | LLM-Finetuning-Toolkit | 676 |
10 | godot-dodo | 511 |
11 | finetune-gpt2xl | 422 |
12 | llm-toys | 115 |
13 | praetor-data | 63 |
14 | nanoChatGPT | 49 |
15 | Zicklein | 32 |
16 | reddit-finetune-frontend | 1 |
Sponsored