-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
A: If you're not already familiar with LLaMA or another local LLM, text generation web UI is currently the best way for running these local models. It has a one-click installer for easy installation and is seeking to be the A1111 of text generation.
LLaMA : https://github.com/facebookresearch/llama And Lit version: https://github.com/Lightning-AI/lit-llamaYou can also use GPTQ-style int4 quantization, but this needs conversions of the weights first bringing GPU usage to about 5GB.
LLaMA : https://github.com/facebookresearch/llama And Lit version: https://github.com/Lightning-AI/lit-llamaYou can also use GPTQ-style int4 quantization, but this needs conversions of the weights first bringing GPU usage to about 5GB.
Related posts
-
Stable LM 3B: Bringing Sustainable, High-Performance LMs to Smart Devices
-
FreeWilly 1 and 2, two new open-access LLMs
-
Losing the plot
-
UAE's Technology Innovation Institute Launches Open-Source "Falcon 40B" Large Language Model for Research & Commercial Utilization
-
Google "We Have No Moat, And Neither Does OpenAI"