-
petals
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
-
Scout Monitoring
Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
https://github.com/bigscience-workshop/petals
Inference only: https://lite.koboldai.net/
If you want to host for the AI Horde, Aphrodite has the most GPU throughput, and Koboldcpp is the easiest and quite flexible:
https://github.com/PygmalionAI/aphrodite-engine
https://github.com/LostRuins/koboldcpp
I havent messed with Petals yet either, but swarm finetuning is very appealing... If thats even implemented atm.
NOTE:
The number of mentions on this list indicates mentions on common posts plus user suggested alternatives.
Hence, a higher number means a more popular project.