-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
This repo, albeit "old" in regards to how much progress there's been in LLMs, has great simple tutorials right there eg. fine-tuning GPT2 with Shakespeare: https://github.com/karpathy/nanoGPT
https://github.com/ggerganov/llama.cpp/pull/1684
You are broadly better off running a larger, more quantized model, than a smaller model at a higher precision. A 65b parameter model with 2bit quantization still outperforms a 30b parameter model in perplexity.
People train LoRAs at low precision regularly. The folks on /r/localllama are really pushing the boundaries of what can be done on consumer hardware, and the boundaries are further out than you'd think.
Related posts
-
Show HN: FileKitty – Combine and label text files for LLM prompt contexts
-
Ask HN: Freelancer? Seeking freelancer? (May 2024)
-
More Low-Bit LLMs
-
Kolmogorov-Arnold Network for Reinforcement Leaning, Initial Experiments
-
Create an AI prototyping environment using Jupyter Lab IDE with Typescript, LangChain.js and Ollama for rapid AI prototyping