Shuffle the Context: RoPE-Perturbed Self-Distillation for Long-Context Adaptation
arXiv:2604.14339v1 Announce Type: new
Abstract: Large language models (LLMs) increasingly operate in settings that require reliable long-context understanding, such as retrieval-augmented generation and multi-document reasoning. A common strategy is t…