Why Does Self-Distillation (Sometimes) Degrade the Reasoning Capability of LLMs?
arXiv:2603.24472v1 Announce Type: cross
Abstract: Self-distillation has emerged as an effective post-training paradigm for LLMs, often improving performance while shortening reasoning traces. However, in mathematical reasoning, we find that it can red…