Memory Inception: Latent-Space KV Cache Manipulation for Steering LLMs
arXiv:2605.06225v1 Announce Type: cross
Abstract: Steering large language models (LLMs) is usually done by either instruction prompting or activation steering. Prompting often gives strong control, but caches guidance tokens at every layer and can clu…