cs.AI, cs.CV

MHSA: A Lightweight Framework for Mitigating Hallucinations via Steered Attention in LVLMs

arXiv:2605.14966v1 Announce Type: cross
Abstract: Large vision-language models (LVLMs) have achieved remarkable performance across diverse multimodal tasks, yet they continue to suffer from hallucinations, generating content that is inconsistent with …