H\’an D\=an Xu\’e B\`u (Mimicry) or Q\=ing Ch\=u Y\’u L\’an (Mastery)? A Cognitive Perspective on Reasoning Distillation in Large Language Models
arXiv:2601.05019v2 Announce Type: replace
Abstract: Recent Large Reasoning Models trained via reinforcement learning exhibit a “natural” alignment with human cognitive costs. However, we show that the prevailing paradigm of reasoning distillation — t…