Online Self-Calibration Against Hallucination in Vision-Language Models
arXiv:2605.00323v1 Announce Type: new
Abstract: Large Vision-Language Models (LVLMs) often suffer from hallucinations, generating descriptions that include visual details absent from the input image. Recent preference alignment methods typically rely …