Response Time Enhances Alignment with Heterogeneous Preferences
arXiv:2605.06987v1 Announce Type: new
Abstract: Aligning large language models (LLMs) to human preferences typically relies on aggregating pooled feedback into a single reward model. However, this standard approach assumes that all labelers share the …