SAGE: Sink-Aware Grounded Decoding for Multimodal Hallucination Mitigation
arXiv:2603.27898v1 Announce Type: new
Abstract: Large vision-language models (VLMs) frequently suffer from hallucinations, generating content that is inconsistent with visual inputs. Existing methods typically address this problem through post-hoc fil…