cs.GT, cs.LG, econ.TH, stat.ML

Response Time Enhances Alignment with Heterogeneous Preferences

arXiv:2605.06987v1 Announce Type: new
Abstract: Aligning large language models (LLMs) to human preferences typically relies on aggregating pooled feedback into a single reward model. However, this standard approach assumes that all labelers share the …