Reasoning Models Will Sometimes Lie About Their Reasoning
arXiv:2601.07663v4 Announce Type: replace
Abstract: Hint-based faithfulness evaluations have established that Large Reasoning Models (LRMs) may not say what they think: they do not always volunteer information about how key parts of the input (e.g. an…