cs.AI

Confidence-Aware Alignment Makes Reasoning LLMs More Reliable

arXiv:2605.07353v1 Announce Type: new
Abstract: Large reasoning models often reach correct answers through flawed intermediate steps, creating a gap between final accuracy and reasoning reliability. Existing alignment strategies address this with exte…