Mitigating Hallucinations in Large Vision-Language Models without Performance Degradation
arXiv:2604.20366v1 Announce Type: new
Abstract: Large Vision-Language Models (LVLMs) exhibit powerful generative capabilities but frequently produce hallucinations that compromise output reliability. Fine-tuning on annotated data devoid of hallucinati…