When Thoughts Meet Facts: Reusable Reasoning for Long-Context LMs
arXiv:2510.07499v2 Announce Type: replace
Abstract: Recent Long-Context Language Models (LCLMs) can process hundreds of thousands of tokens in a single prompt, enabling new opportunities for knowledge-intensive multi-hop reasoning by integrating large…