My experience with Alpaca.cpp

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • alpaca.cpp

    Discontinued Locally run an Instruction-Tuned Chat-Style LLM

  • I've gotten that error with various attempts at using llama.cpp/alpaca.cpp but honestly don't know what it means. If you Google it, there's not much out there (really just this thread and a couple like it that don't provide much of use). I assume it's some sort of internal "checksum" meant to verify that the model file is indeed a valid model or in the correct format. Did you download the model from the link above? If not try that. If so then I really don't know how to fix it unless there's some new format being used.

  • alpaca-lora

    Instruct-tune LLaMA on consumer hardware

  • In theory, something like this could be used to do it, but according to that source, it took about 5 hours on a 4090 to train the 7B variant even with lora. I've also heard it takes about 18 GB of VRAM to train the 7B variant. Assuming everything scales proportionally, that's ~170 GB to fine-tune the 65B variant. Doing that with 8 A100s for instance would cost a little over $30/hour.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • alpaca_lora_4bit

  • There's a repo for tuning Loras on the 4-bit models. Readme says it can train 30B on a single 24GB card with Gradient Checkpointing enabled (which does slow things down quite a lot).

  • llamacpp-for-kobold

    Discontinued Port of Facebook's LLaMA model in C/C++ [Moved to: https://github.com/LostRuins/koboldcpp]

  • I don't know if anything like that exists. There is this project that I played around with at one point if that helps at all.

  • alpaca.cpp

    Locally run an Instruction-Tuned Chat-Style LLM (by trevtravtrev)

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts