ParoQuant: Pairwise Rotation Quantization for Efficient Reasoning LLM Inference
Abstract
Pairwise Rotation Quantization (ParoQuant) improves the accuracy of quantized Large Language Models by suppressing outliers and reducing quantization errors through independent Givens rotations and channel-wise scaling, while maintaining low inference overhead.
Weight-only post-training quantization (PTQ) compresses the weights of Large Language Models (LLMs) into low-precision representations to reduce memory footprint and accelerate inference. However, the presence of outliers in weights and activations often leads to large quantization errors and severe accuracy degradation, especially in recent reasoning LLMs where errors accumulate across long chains of thought. Existing PTQ methods either fail to sufficiently suppress outliers or introduce significant overhead during inference. In this paper, we propose Pairwise Rotation Quantization (ParoQuant), a weight-only PTQ method that combines hardware-efficient and optimizable independent Givens rotations with channel-wise scaling to even out the magnitude across channels and narrow the dynamic range within each quantization group. We further co-design the inference kernel to fully exploit GPU parallelism and keep the rotations and scaling lightweight at runtime. ParoQuant achieves an average 2.4% accuracy improvement over AWQ on reasoning tasks with less than 10% overhead. This paves the way for more efficient and accurate deployment of reasoning LLMs.
Community
arXiv explained breakdown of this paper ๐ https://arxivexplained.com/papers/paroquant-pairwise-rotation-quantization-for-efficient-reasoning-llm-inference
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper