Think Through Uncertainty: Improving Long-Form Generation Factuality via Reasoning Calibration
arXiv:2604.12046v1 Announce Type: new
Abstract: Large language models (LLMs) often hallucinate in long-form generation. Existing approaches mainly improve factuality through post-hoc revision or reinforcement learning (RL) with correctness-based rewar…