-
Scout Monitoring
Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
-
Open-Instructions
Open-Instructions: A Pavilion of recent Open Source GPT Projects for decentralized AI.
-
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Take a look at this: https://github.com/qwopqwop200/GPTQ-for-LLaMa
This project implements using c++ instead of Python for performance optimizations with a focus on CPU only systems: https://github.com/ggerganov/llama.cpp They use quantization compression to dramatically shrink the size of the model so that it will fit in limited RAM capacities. Many existing models have already been converted to be compatible with llama.cpp but more recent ones (like Dolly 2.0) may still need to be converted. The project provides tools and scripts to make it easier for users to convert and/or quantize models into a format compatible with llama.cpp.
We release an open-source project named Open-Instructions to help the community gather all the recently released datasets for instruction finetuning, with format already been converted to conversations so compatible with Vicuna training pipeline. And you can train LLaMA using Dolly's real-world data rather than only gpt turbo, can't wait to see the performance.
I'll be also putting together a Dolly 2.0 series here: https://github.com/kw2828/Dolly-2.0-Series