Detecting Hallucinations in Large Language Models via Internal Attention Divergence Signals
arXiv:2605.05025v1 Announce Type: new
Abstract: We propose a lightweight and single-pass uncertainty quantification method for detecting hallucinations in Large Language Models. The method uses attention matrices to estimate uncertainty without requir…