-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Thank you for bringing that to my attention ! I can't (without starving to death) spend more than around 100 until i can afford another real computer. I guess i'll poke around and check anyway this part about "docker". However i'll need to poke around since : https://github.com/oobabooga/text-generation-webui Mention that i should be using " TORCH_CUDA_ARCH_LIST" Based on my gpu and i have no knowledge what is the replacement for my poor's man GPU intel graphic.
I run vicuna-7b in browser on my MacBook Pro M1 via https://github.com/mlc-ai/mlc-llm
As far as I know, you only need a single ggml .bin file for CPU inference. I use koboldcpp and it's just drag&drop .bin on top of .exe to make it work.
Of course, the bigger the model, the longer it takes. 7B q5_1 generations take about 400-450 ms/Token, 13B q5_1 about 700-800 ms/T. Thanks to a flood of optimizations, things have been improving steadily, and stuff like Proof of concept: GPU-accelerated token generation will soon provide another much needed and welcome boost.