KoCo: Conditioning Language Model Pre-training on Knowledge Coordinates
arXiv:2604.12397v1 Announce Type: new
Abstract: Standard Large Language Model (LLM) pre-training typically treats corpora as flattened token sequences, often overlooking the real-world context that humans naturally rely on to contextualize information…