-
alpaca-electron
The simplest way to run Alpaca (and other LLaMA-based local LLMs) on your own computer
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
4GB of VRAM is enough to run a 7B model. Try this, https://github.com/mlc-ai/mlc-llm. It uses Vulkan instead of CUDA. Their converted 7B model runs on a GTX 4GB card for me. It's pretty speedy too.
They work great for me. No need to get fancy, the easiest thing is to use the distribution from the person who wrote the code that everyone else is using, it works great. It's at https://github.com/ggerganov/llama.cpp. I use it with everything from 7B to 65B models.