RecaLLM: Addressing the Lost-in-Thought Phenomenon with Explicit In-Context Retrieval
arXiv:2604.09494v1 Announce Type: cross
Abstract: We propose RecaLLM, a set of reasoning language models post-trained to make effective use of long-context information. In-context retrieval, which identifies relevant evidence from context, and reasoni…