Towards Understanding Continual Factual Knowledge Acquisition of Language Models: From Theory to Algorithm
arXiv:2605.10640v1 Announce Type: cross
Abstract: Continual Pre-Training (CPT) is essential for enabling Language Models (LMs) to integrate new knowledge without erasing old. While classical CPT techniques like data replay have become the standard par…