[D] Have you tried fine-tuning an open source LLM?

This page summarizes the projects mentioned and recommended in the original post on /r/MachineLearning

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • alpaca-lora

    Instruct-tune LLaMA on consumer hardware

  • Using https://github.com/tloen/alpaca-lora I can train a LLaMA 7B on 3 epochs on my own dual 3090 cards for 15 hours for around 80 cents of electricity. So I can kick off a train at 5pm, call it day, then at 8am the next day I'll have a testable train I can play with for less than a dollar. Odds are the model may not work the way I want, or maybe I have a better idea on how to improve the data set and so on.

  • llm-foundry

    LLM training code for Databricks foundation models

  • I'm trying to run fine tuning on an A10G video card but keep running into out of memory errors, even with the default settings provided by their examples -- for example, this one here: https://github.com/mosaicml/llm-foundry/blob/main/scripts/train/yamls/finetune/mpt-7b_dolly_sft.yaml

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • LMFlow

    An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

  • I'd like to recommend LMFlow (https://github.com/OptimalScale/LMFlow), a fast and extensible toolkit for finetuning and inference of large foundation models.

  • transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

  • Currently we are using a GPT-2 style model that is ~ 1 B in params. This model can be fine tuned on Nvidia Jetson Xavier device. reComputer from SeedStudio can work for this - https://www.seeedstudio.com/reComputer-J2022-p-5497.html fine tuning is implemented using the standard script from hugging face - https://github.com/huggingface/transformers/tree/main/examples/pytorch/language-modeling

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Database of 16,000 Artists Used to Train Midjourney AI Goes Viral

    1 project | news.ycombinator.com | 7 Jan 2024
  • People who've used RWKV, whats your wishlist for it?

    9 projects | /r/LocalLLaMA | 9 Dec 2023
  • A look at Apple’s new Transformer-powered predictive text model

    4 projects | news.ycombinator.com | 16 Sep 2023
  • Train Your AI Model Once and Deploy on Any Cloud

    3 projects | news.ycombinator.com | 8 Jul 2023
  • Your weekly machine learning digest

    2 projects | /r/learnmachinelearning | 3 Jul 2023