Correcting Influence: Unboxing LLM Outputs with Orthogonal Latent Spaces
arXiv:2605.12809v1 Announce Type: new
Abstract: A critical step for reliable large language models (LLMs) use in healthcare is to attribute predictions to their training data, akin to a medical case study. This requires token-level precision: pinpoint…