cs.CV, cs.LG

Improving Diversity in Black-box Few-shot Knowledge Distillation

arXiv:2604.25795v1 Announce Type: cross
Abstract: Knowledge distillation (KD) is a well-known technique to effectively compress a large network (teacher) to a smaller network (student) with little sacrifice in performance. However, most KD methods req…