Risk Awareness Injection: Calibrating Vision-Language Models for Safety without Compromising Utility
arXiv:2602.03402v3 Announce Type: replace
Abstract: Vision language models (VLMs) extend the reasoning capabilities of large language models (LLMs) to cross-modal settings, yet remain highly vulnerable to multimodal jailbreak attacks. Existing defense…