Route Before Retrieve: Activating Latent Routing Abilities of LLMs for RAG vs. Long-Context Selection
arXiv:2605.10235v1 Announce Type: new
Abstract: Recent advances in large language models (LLMs) have expanded the context window to beyond 128K tokens, enabling long-document understanding and multi-source reasoning. A key challenge, however, lies in …