From ad442a22b5cfcd223d16ad6d20d0ad617c2a217b Mon Sep 17 00:00:00 2001 From: Casper Date: Fri, 21 Jun 2024 10:19:37 +0200 Subject: [PATCH] Update RunPod support (#514) --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 13dca0b7..e83ebb56 100644 --- a/README.md +++ b/README.md @@ -18,8 +18,8 @@

Supported by

- - RunPod Logo + + RunPod Logo
AutoAWQ is an easy-to-use package for 4-bit quantized models. AutoAWQ speeds up models by 3x and reduces memory requirements by 3x compared to FP16. AutoAWQ implements the Activation-aware Weight Quantization (AWQ) algorithm for quantizing LLMs. AutoAWQ was created and improved upon from the [original work](https://github.com/mit-han-lab/llm-awq) from MIT.