[P] Finetuning a commercially viable open source LLM (Flan-UL2) using Alpaca, Dolly15K and LoRA

This page summarizes the projects mentioned and recommended in the original post on /r/MachineLearning

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
  • alpaca-lora

    Instruct-tune LLaMA on consumer hardware

  • ue5-llama-lora

    A proof-of-concept project that showcases the potential for using small, locally trainable LLMs to create next-generation documentation tools.

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • dolly

    Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform

  • Hi /u/meowkittykitty510, I'm embarking on a similar project, solely for teaching myself about finetuning LLMs. I initially chose Dolly (Pythia model) with modifications to the Dolly 2.0 trainer plus additional training data scraped from the web about The Expanse just as a fun way to test.

  • AlpacaDataCleaned

    Alpaca dataset from Stanford, cleaned and curated

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts