Forget What Matters, Keep the Rest: Selective Unlearning of Informative Tokens
arXiv:2604.17785v1 Announce Type: new
Abstract: Unlearning in large language models (LLMs) has emerged as a promising safeguard against adversarial behaviors. When the forgetting loss is applied uniformly without considering token-level semantic impor…