SaaSHub helps you find the best software and product alternatives Learn more →
Top 15 Python instruction-tuning Projects
-
Project mention: Llama-Factory: A WebUI for Efficient Fine-Tuning of 100 LLMs | news.ycombinator.com | 2024-07-17
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Project mention: Show HN: LLM Aided OCR (Correcting Tesseract OCR Errors with LLMs) | news.ycombinator.com | 2024-08-09This package seems to use llama_cpp for local inference [1] so you can probably use anything supported by that [2]. However, I think it's just passing OCR output for correction - the language model doesn't actually see the original image.
That said, there are some large language models you can run locally which accept image input. Phi-3-Vision [3], LLaVA [4], MiniCPM-V [5], etc.
[1] - https://github.com/Dicklesworthstone/llm_aided_ocr/blob/main...
[2] - https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#de...
[3] - https://huggingface.co/microsoft/Phi-3-vision-128k-instruct
[4] - https://github.com/haotian-liu/LLaVA
[5] - https://github.com/OpenBMB/MiniCPM-V
-
Here’s another one - it’s older but has some interesting charts and graphs.
https://arxiv.org/abs/2303.18223
-
-
Otter
🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
-
-
Video-LLaVA
【EMNLP 2024🔥】Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
-
-
Project mention: Cambrian-1 an Open, Vision-Centric Exploration of Multimodal LLMs | news.ycombinator.com | 2024-06-25
Code: [cambrian-mllm/*cambrian*](https://github.com/cambrian-mllm/cambrian)
-
-
-
DoRA
[ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed Low-Rank Adaptation (by NVlabs)
-
HugNLP
CIKM2023 Best Demo Paper Award. HugNLP is a unified and comprehensive NLP library based on HuggingFace Transformer. Please hugging for NLP now!😊 (by HugAILab)
-
-
Python instruction-tuning discussion
Python instruction-tuning related posts
-
Google Bard AI Now Has the Ability to Understand YouTube Videos
-
Video-LLaVA
-
Video-LLaVA
-
Share your favorite materials: intersection of LLMs and business applications
-
Recommended open LLMs with image input modality?
-
HugNLP: A Unified and Comprehensive Open-Source Library for NLP
-
[R] CodeCapybara: Another open source model for code generation based on instruction tuning, outperformed Llama and CodeAlpaca
-
A note from our sponsor - SaaSHub
www.saashub.com | 12 Oct 2024
Index
What are some of the best open-source instruction-tuning projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | LLaMA-Factory | 31,926 |
2 | LLaVA | 19,655 |
3 | LLMSurvey | 10,139 |
4 | self-instruct | 4,080 |
5 | Otter | 3,560 |
6 | NExT-GPT | 3,241 |
7 | Video-LLaVA | 2,888 |
8 | mPLUG-Owl | 2,271 |
9 | cambrian | 1,713 |
10 | InternVideo | 1,338 |
11 | DataDreamer | 813 |
12 | DoRA | 591 |
13 | HugNLP | 374 |
14 | CodeCapybara | 159 |
15 | tasksource | 144 |