cs.CL, cs.CR

ContextualJailbreak: Evolutionary Red-Teaming via Simulated Conversational Priming

arXiv:2605.02647v1 Announce Type: new
Abstract: Large language models (LLMs) remain vulnerable to jailbreak attacks that bypass safety alignment and elicit harmful responses. A growing body of work shows that contextual priming, where earlier turns co…