cs.AI, cs.CR

SALLIE: Safeguarding Against Latent Language & Image Exploits

arXiv:2604.06247v1 Announce Type: cross
Abstract: Large Language Models (LLMs) and Vision-Language Models (VLMs) remain highly vulnerable to textual and visual jailbreaks, as well as prompt injections (arXiv:2307.15043, Greshake et al., 2023, arXiv:23…