cs.AI, cs.CR

VisInject: Disruption != Injection — A Dual-Dimension Evaluation of Universal Adversarial Attacks on Vision-Language Models

arXiv:2605.01449v1 Announce Type: cross
Abstract: Universal adversarial attacks on aligned multimodal large language models are increasingly reported with attack success rates in the 60-80% range, suggesting the visual modality is highly vulnerable to…