Probing Persona-Dependent Preferences in Language Models
arXiv:2605.13339v1 Announce Type: cross
Abstract: Large language models (LLMs) can be said to have preferences: they reliably pick certain tasks and outputs over others, and preferences shaped by post-training and system prompts appear to shape much o…