-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
llamacpp-for-kobold
Discontinued Port of Facebook's LLaMA model in C/C++ [Moved to: https://github.com/LostRuins/koboldcpp]
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
I've gotten that error with various attempts at using llama.cpp/alpaca.cpp but honestly don't know what it means. If you Google it, there's not much out there (really just this thread and a couple like it that don't provide much of use). I assume it's some sort of internal "checksum" meant to verify that the model file is indeed a valid model or in the correct format. Did you download the model from the link above? If not try that. If so then I really don't know how to fix it unless there's some new format being used.
In theory, something like this could be used to do it, but according to that source, it took about 5 hours on a 4090 to train the 7B variant even with lora. I've also heard it takes about 18 GB of VRAM to train the 7B variant. Assuming everything scales proportionally, that's ~170 GB to fine-tune the 65B variant. Doing that with 8 A100s for instance would cost a little over $30/hour.
There's a repo for tuning Loras on the 4-bit models. Readme says it can train 30B on a single 24GB card with Gradient Checkpointing enabled (which does slow things down quite a lot).
I don't know if anything like that exists. There is this project that I played around with at one point if that helps at all.