cs.CL

EGAD: Entropy-Guided Adaptive Distillation for Token-Level Knowledge Transfer

arXiv:2605.01732v1 Announce Type: new
Abstract: Large language models (LLMs) have achieved remarkable performance across diverse domains, yet their enormous computational and memory requirements hinder deployment in resource-constrained environments. …