-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Using https://github.com/tloen/alpaca-lora I can train a LLaMA 7B on 3 epochs on my own dual 3090 cards for 15 hours for around 80 cents of electricity. So I can kick off a train at 5pm, call it day, then at 8am the next day I'll have a testable train I can play with for less than a dollar. Odds are the model may not work the way I want, or maybe I have a better idea on how to improve the data set and so on.
I'm trying to run fine tuning on an A10G video card but keep running into out of memory errors, even with the default settings provided by their examples -- for example, this one here: https://github.com/mosaicml/llm-foundry/blob/main/scripts/train/yamls/finetune/mpt-7b_dolly_sft.yaml
I'd like to recommend LMFlow (https://github.com/OptimalScale/LMFlow), a fast and extensible toolkit for finetuning and inference of large foundation models.
Currently we are using a GPT-2 style model that is ~ 1 B in params. This model can be fine tuned on Nvidia Jetson Xavier device. reComputer from SeedStudio can work for this - https://www.seeedstudio.com/reComputer-J2022-p-5497.html fine tuning is implemented using the standard script from hugging face - https://github.com/huggingface/transformers/tree/main/examples/pytorch/language-modeling