RouteHijack: Routing-Aware Attack on Mixture-of-Experts LLMs
arXiv:2605.02946v1 Announce Type: new
Abstract: Safety alignment is critical for the responsible deployment of large language models (LLMs). As Mixture-of-Experts (MoE) architectures are increasingly adopted to scale model capacity, understanding thei…