cs.AI

How LLMs Are Persuaded: A Few Attention Heads, Rerouted

arXiv:2605.09314v1 Announce Type: new
Abstract: Language models can be persuaded to abandon factual knowledge. This vulnerability is central to AI safety, but its internal mechanism remains poorly understood. We uncover a compact causal mechanism for …