cs.LG

When Flat Minima Fail: Characterizing INT4 Quantization Collapse After FP32 Convergence

arXiv:2604.15167v1 Announce Type: new
Abstract: Post-training quantization (PTQ) assumes that a well-converged model is a quantization-ready model. We show this assumption fails in a structured, measurable, and previously uncharacterized way. Using a …