cs.AI, cs.CL

One Adapts to Any: Meta Reward Modeling for Personalized LLM Alignment

arXiv:2601.18731v2 Announce Type: replace
Abstract: Alignment of Large Language Models (LLMs) aims to align outputs with human preferences, and personalized alignment further adapts models to individual users. This relies on personalized reward models…