Make Your LVLM KV Cache More Lightweight
arXiv:2605.00789v1 Announce Type: cross
Abstract: Key-Value (KV) cache has become a de facto component of modern Large Vision-Language Models (LVLMs) for inference. While it enhances decoding efficiency in Large Language Models (LLMs), its direct adop…