cs.AI, cs.CR

Enhancing Jailbreak Attacks on LLMs via Persona Prompts

arXiv:2507.22171v3 Announce Type: replace-cross
Abstract: Jailbreak attacks aim to exploit large language models (LLMs) by inducing them to generate harmful content, thereby revealing their vulnerabilities. Understanding and addressing these attacks i…