cs.LG

LoRaQ: Optimized Low Rank Approximation for 4-bit Quantization

arXiv:2604.18117v1 Announce Type: new
Abstract: Post-training quantization (PTQ) is essential for deploying large diffusion transformers on resource-constrained hardware, but aggressive 4-bit quantization significantly degrades generative performance….