cs.AI, cs.CL

Hallucination as output-boundary misclassification: a composite abstention architecture for language models

arXiv:2604.06195v1 Announce Type: cross
Abstract: Large language models often produce unsupported claims. We frame this as a misclassification error at the output boundary, where internally generated completions are emitted as if they were grounded in…