cs.LG

OSAQ: Outlier Self-Absorption for Accurate Low-bit LLM Quantization

arXiv:2605.04738v1 Announce Type: new
Abstract: Large Language Models (LLMs) have demonstrated remarkable capabilities. However, their massive parameter scale leads to significant resource consumption and latency during inference. Post-training weight…