cs.AI, cs.CL

Sanity Checks for Long-Form Hallucination Detection

arXiv:2605.08346v1 Announce Type: cross
Abstract: Hallucination detection methods for large language models increasingly operate on chain-of-thought reasoning traces, yet it remains unclear whether they evaluate the reasoning itself or merely exploit …