Latent Personality Alignment: Improving Harmlessness Without Mentioning Harms
arXiv:2605.08496v1 Announce Type: new
Abstract: Current adversarial robustness methods for large language models require extensive datasets of harmful prompts (thousands to hundreds of thousands of examples), yet remain vulnerable to novel attack vect…