Tell Model Where to Look: Mitigating Hallucinations in MLLMs by Vision-Guided Attention
arXiv:2511.20032v3 Announce Type: replace
Abstract: Visual attention serves as the primary mechanism through which MLLMs interpret visual information; however, its limited localization capability often leads to hallucinations. We observe that although…