cs.AI, cs.LG

Robust Multimodal Safety via Conditional Decoding

arXiv:2604.00310v1 Announce Type: cross
Abstract: Multimodal large-language models (MLLMs) often experience degraded safety alignment when harmful queries exploit cross-modal interactions. Models aligned on text alone show a higher rate of successful …

Scroll to Top