PolarQuant: Optimal Gaussian Weight Quantization via Hadamard Rotation for LLM Compression
arXiv:2603.29078v2 Announce Type: replace-cross
Abstract: We present PolarQuant, a post-training weight quantization method for large language models (LLMs) that exploits the distributional structure of neural network weights to achieve near-lossless …