cs.CL

Reducing Hallucinations in LLMs via Factuality-Aware Preference Learning

arXiv:2601.03027v3 Announce Type: replace
Abstract: Preference alignment methods such as RLHF and Direct Preference Optimization (DPO) improve instruction following, but they can also reinforce hallucinations when preference judgments reward fluency a…