cs.LG

On the Blessing of Pre-training in Weak-to-Strong Generalization

arXiv:2605.05710v1 Announce Type: new
Abstract: The paradigm of Weak-to-Strong Generalization (W2SG) suggests that a pre-trained strong model can surpass its weak supervisor, yet the decisive role of pre-training remains theoretically and empirically …