Sanity Checks for Long-Form Hallucination Detection
arXiv:2605.08346v1 Announce Type: cross
Abstract: Hallucination detection methods for large language models increasingly operate on chain-of-thought reasoning traces, yet it remains unclear whether they evaluate the reasoning itself or merely exploit …