cs.AI, cs.CR

Benchmarking Safety Risks of Knowledge-Intensive Reasoning under Malicious Knowledge Editing

arXiv:2605.10146v1 Announce Type: new
Abstract: Large language models (LLMs) increasingly rely on knowledge editing to support knowledge-intensive reasoning, but this flexibility also introduces critical safety risks: adversaries can inject malicious …