cs.AI, cs.LG

Curriculum Learning-Guided Progressive Distillation in Large Language Models

arXiv:2605.11260v1 Announce Type: new
Abstract: Knowledge distillation is a key technique for transferring the capabilities of large language models (LLMs) into smaller, more efficient student models. Existing distillation approaches often overlook tw…