cs.CL, cs.LG

Attention Sinks as Internal Signals for Hallucination Detection in Large Language Models

arXiv:2604.10697v1 Announce Type: new
Abstract: Large language models frequently exhibit hallucinations: fluent and confident outputs that are factually incorrect or unsupported by the input context. While recent hallucination detection methods have e…