cs.AI, cs.CL, cs.LG

Where does output diversity collapse in post-training?

arXiv:2604.16027v1 Announce Type: cross
Abstract: Post-trained language models produce less varied outputs than their base counterparts. This output diversity collapse undermines inference-time scaling methods that rely on varied samples, and risks ho…