How can I train my custom dataset on top of Vicuna?

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • alpaca-lora

    Instruct-tune LLaMA on consumer hardware

  • LLaMA-LoRA-Tuner

    UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT-like Chat UI to demonstrate your language models.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • stanford_alpaca

    Code and documentation to train Stanford's Alpaca models, and generate the data.

  • If you have access to the data-center grade GPU, the quickest way to start would be to pick one of the efforts of fine-tuning, for example, stanford alpaca (https://github.com/tatsu-lab/stanford_alpaca/ ) or indeed, vicuna (https://github.com/lm-sys/FastChat ) and use your own data. The main issue for home users is that their VRAM is vastly insufficient for standard model tuning (original weights, updated weights, a copy of adam himself, and a copy for the AI overlord…)

  • FastChat

    An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.

  • If you have access to the data-center grade GPU, the quickest way to start would be to pick one of the efforts of fine-tuning, for example, stanford alpaca (https://github.com/tatsu-lab/stanford_alpaca/ ) or indeed, vicuna (https://github.com/lm-sys/FastChat ) and use your own data. The main issue for home users is that their VRAM is vastly insufficient for standard model tuning (original weights, updated weights, a copy of adam himself, and a copy for the AI overlord…)

  • simple-llm-finetuner

    Discontinued Simple UI for LLM Model Finetuning

  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

  • Sure! This is the link to Oobabooga https://github.com/oobabooga/text-generation-webui

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • [P] Uptraining a pretrained model using company data?

    4 projects | /r/MachineLearning | 25 May 2023
  • (HELP) Token Issue on Generation

    1 project | /r/LocalLLaMA | 19 May 2023
  • Help with Random Characters and Words on Output

    1 project | /r/LocalLLaMA | 18 May 2023
  • Fine-tuning LLaMA for research without Meta license

    1 project | /r/LocalLLaMA | 15 May 2023
  • Why run LLMs locally?

    4 projects | /r/LocalLLaMA | 8 May 2023