ContextualJailbreak: Evolutionary Red-Teaming via Simulated Conversational Priming
arXiv:2605.02647v1 Announce Type: new
Abstract: Large language models (LLMs) remain vulnerable to jailbreak attacks that bypass safety alignment and elicit harmful responses. A growing body of work shows that contextual priming, where earlier turns co…