Our great sponsors
-
llamacpp-for-kobold
Discontinued Port of Facebook's LLaMA model in C/C++ [Moved to: https://github.com/LostRuins/koboldcpp]
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
Alpaca.cpp is extremely simple to get up and running. You don't need any Conda environments, don't need to install Linux or WSL, don't need to install Python, CUDA, anything at all. It's a single ~200kb EXE that you just run, and you put a 4GB model file into the directory. That's it.
Try this https://github.com/LostRuins/llamacpp-for-kobold
https://github.com/oobabooga/text-generation-webui/pull/447 this, in theory, could easily run Alpaca via C++ exe backend and oobabooga webUI interface