Fewer Weights, More Problems: A Practical Attack on LLM Pruning
arXiv:2510.07985v3 Announce Type: replace-cross
Abstract: Model pruning, i.e., removing a subset of model weights, has become a prominent approach to reducing the memory footprint of large language models (LLMs) during inference. Notably, popular infe…