cs.CL

Once Correct, Still Wrong: Counterfactual Hallucination in Multilingual Vision-Language Models

arXiv:2602.05437v2 Announce Type: replace
Abstract: Vision-language models (VLMs) can achieve high accuracy while still accepting culturally plausible but visually incorrect interpretations. Existing hallucination benchmarks rarely test this failure m…