Principled Detection of Hallucinations in Large Language Models via Multiple Testing
arXiv:2508.18473v3 Announce Type: replace
Abstract: While Large Language Models (LLMs) have emerged as powerful foundational models to solve a variety of tasks, they have also been shown to be prone to hallucinations, i.e., generating responses that s…