Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
SqueezeLLM Alternatives
Similar projects and alternatives to SqueezeLLM
-
Qwen-7B
Discontinued The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud. [Moved to: https://github.com/QwenLM/Qwen]
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
Qwen
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
-
Pretrained-Language-Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
SqueezeLLM reviews and mentions
-
Llama33B vs Falcon40B vs MPT30B
Using the currently popular gptq the 3bit quantization hurts performance much more than 4bit, but there's also awq (https://github.com/mit-han-lab/llm-awq) and squishllm (https://github.com/SqueezeAILab/SqueezeLLM) which are able to manage 3bit without as much performance drop - I hope to see them used more commonly.
-
Has anyone tried out Squeezellm?
[Paper][Github][Model]
- SqueezeLLM: Dense-and-Sparse Quantization
- New quantization method SqueezeLLM allows for loseless compression for 3-bit and outperforms GPTQ and AWQ in both 3-bit and 4-bit. Quantized Vicuna and LLaMA models have been released.
-
A note from our sponsor - InfluxDB
www.influxdata.com | 2 May 2024
Stats
SqueezeAILab/SqueezeLLM is an open source project licensed under MIT License which is an OSI approved license.
The primary programming language of SqueezeLLM is Python.
Sponsored