[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Why do you think that https://github.com/mehulgit/optiagent is a good alternative to llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Why do you think that https://github.com/mehulgit/optiagent is a good alternative to llm-awq