RAG or Learning? Understanding the Limits of LLM Adaptation under Continuous Knowledge Drift in the Real World
arXiv:2604.05096v2 Announce Type: replace
Abstract: Large language models (LLMs) acquire most of their knowledge during pretraining, which ties them to a fixed snapshot of the world and makes adaptation to continuously evolving knowledge challenging. …