A Survey of On-Policy Distillation for Large Language Models
arXiv:2604.00626v1 Announce Type: cross
Abstract: Knowledge distillation has become a primary mechanism for transferring reasoning and domain expertise from frontier Large Language Models (LLMs) to smaller, deployable students. However, the dominant p…