-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
I am the author of pyllama https://github.com/juncongmoo/pyllama
What's with everyone here talking out of their own arse? LLaMA 7B runs on the CPU (I work on llama.rs) or on the GPU.
What's with everyone here talking out of their own arse? LLaMA 7B runs on the CPU (I work on llama.rs) or on the GPU.
What's with everyone here talking out of their own arse? LLaMA 7B runs on the CPU (I work on llama.rs) or on the GPU.
4-bit GPTQ performs extremely well and allows for 7B LLaMA in 4 gigabytes.
Related posts
-
AlphaFold 3 predicts the structure and interactions of all of life's molecules
-
TimesFM (Time Series Foundation Model) for time-series forecasting
-
A library to assist writing memory-unsafe code in "pure" Python
-
2D Gaussian Splatting for Geometrically Accurate Radiance Fields
-
CDK CodeChecker v2: Breaking Free from Third-Party Dependencies