Our great sponsors
-
llamacpp-for-kobold
Discontinued Port of Facebook's LLaMA model in C/C++ [Moved to: https://github.com/LostRuins/koboldcpp]
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
TavernAI
Discontinued TavernAI for nerds [Moved to: https://github.com/Cohee1207/SillyTavern] (by SillyLossy)
There's an important fix for 65B models upstream: https://github.com/ggerganov/llama.cpp/pull/438/files. I've verified it works on my local copy. Can your fork be updated from upstream? Without it llama will segfault to an under-estimate of the memory required.
Enter llamacpp-for-kobold
There's also a single file version, where you just drag-and-drop your llama model onto the .exe file, and connect KoboldAI to the displayed link.
It looks like some endpoints are missing that flavors of TavernAI depend on. E.g., this promising version of TavernAI needs /config/soft_prompts