cs.LG

Resting Neurons, Active Insights: Robustify Activation Sparsity for Large Language Models

arXiv:2512.12744v2 Announce Type: replace
Abstract: Activation sparsity offers a compelling route to accelerate large language model (LLM) inference by selectively suppressing hidden activations, yet existing approaches exhibit severe accuracy degrada…