Optimizing Korean-Centric LLMs via Token Pruning
arXiv:2604.16235v1 Announce Type: new
Abstract: This paper presents a systematic benchmark of state-of-the-art multilingual large language models (LLMs) adapted via token pruning – a compression technique that eliminates tokens and embedding parameter…