Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Top 23 Python fine-tuning Projects
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
xTuring
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
-
custom-diffusion
Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
-
Lora-for-Diffusers
The most easy-to-understand tutorial for using LoRA (Low-Rank Adaptation) within diffusers framework for AI Generation Researchers🔥
-
simpleT5
simpleT5 is built on top of PyTorch-lightning⚡️ and Transformers🤗 that lets you quickly train your T5 models.
-
discus
A data-centric AI package for ML/AI. Get the best high-quality data for the best results. Discord: https://discord.gg/t6ADqBKrdZ
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
Project mention: LlamaIndex: A data framework for your LLM applications | news.ycombinator.com | 2024-04-07
Depends what model you want to train, and how well you want your computer to keep working while you're doing it.
If you're interested in large language models there's a table of vram requirements for fine-tuning at [1] which says you could do the most basic type of fine-tuning on a 7B parameter model with 8GB VRAM.
You'll find that training takes quite a long time, and as a lot of the GPU power is going on training, your computer's responsiveness will suffer - even basic things like scrolling in your web browser or changing tabs uses the GPU, after all.
Spend a bit more and you'll probably have a better time.
[1] https://github.com/hiyouga/LLaMA-Factory?tab=readme-ov-file#...
Project mention: Show HN: Toolkit for LLM Fine-Tuning, Ablating and Testing | news.ycombinator.com | 2024-04-07This is a great project, little bit similar to https://github.com/ludwig-ai/ludwig, but it includes testing capabilities and ablation.
questions regarding the LLM testing aspect: How extensive is the test coverage for LLM use cases, and what is the current state of this project area? Do you offer any guarantees, or is it considered an open-ended problem?
Would love to see more progress toward this area!
Project mention: I'm developing an open-source AI tool called xTuring, enabling anyone to construct a Language Model with just 5 lines of code. I'd love to hear your thoughts! | /r/machinelearningnews | 2023-09-07Explore the project on GitHub here.
Project mention: YiVal——Unlocking Your Data's Power to Create Customized GenAI Apps | /r/u_YiVal | 2023-11-16- 🤖Github:https://github.com/YiVal/YiVal/pull/189
Project mention: Now You Can Full Fine Tune / DreamBooth Stable Diffusion XL (SDXL) with only 10.3 GB VRAM via OneTrainer | dev.to | 2024-03-25Used SG161222/RealVisXL_V4.0 as a base model and OneTrainer to train on Windows 10 : https://github.com/Nerogar/OneTrainer
Project mention: Google DeepMind CEO Says Some Form of AGI Possible in a Few Years | /r/singularity | 2023-05-03That is not true, you can for example use an additional adapter to optimize, that takes 50$ and a 1 hour. https://github.com/AGI-Edgerunners/LLM-Adapters
Project mention: Show HN: Toolkit for LLM Fine-Tuning, Ablating and Testing | news.ycombinator.com | 2024-04-07
slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization [Link].
Project mention: 25 million Creative Commons image dataset released! | /r/StableDiffusion | 2023-10-01Github: https://github.com/ml6team/fondant
Project mention: an open source package helping developers generate data for LLMs | /r/mlops | 2023-08-02
Python fine-tuning related posts
- penzai: JAX research toolkit for building, editing, and visualizing neural nets
- Show HN: Toolkit for LLM Fine-Tuning, Ablating and Testing
- LlamaIndex: A data framework for your LLM applications
- LlamaIndex is a data framework for your LLM applications
- Now You Can Full Fine Tune / DreamBooth Stable Diffusion XL (SDXL) with only 10.3 GB VRAM via OneTrainer
- Geniusrise – Wannabe Competitor to Vertex AI, Azure AI Studio and Bedrock
- Show HN: GPU Prices on eBay
-
A note from our sponsor - InfluxDB
www.influxdata.com | 23 Apr 2024
Index
What are some of the best open-source fine-tuning projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | llama_index | 30,910 |
2 | LLaMA-Factory | 17,050 |
3 | ludwig | 10,778 |
4 | xTuring | 2,515 |
5 | YiVal | 2,425 |
6 | custom-diffusion | 1,776 |
7 | finetuner | 1,423 |
8 | OneTrainer | 1,076 |
9 | TencentPretrain | 975 |
10 | LLM-Adapters | 936 |
11 | SPIN | 773 |
12 | Lora-for-Diffusers | 696 |
13 | LLM-Finetuning-Toolkit | 659 |
14 | DataDreamer | 632 |
15 | slowllama | 413 |
16 | simpleT5 | 383 |
17 | fondant | 316 |
18 | OneDiffusion | 315 |
19 | kiri | 240 |
20 | Dreambooth | 94 |
21 | penzai | 72 |
22 | praetor-data | 63 |
23 | discus | 62 |
Sponsored