Stabilizing LLM Supervised Fine-Tuning via Explicit Distributional Control
arXiv:2605.04468v1 Announce Type: new
Abstract: Post-training large language models (LLMs) often suffers from catastrophic forgetting, where improvements on a target objective degrade previously acquired capabilities. Recent evidence suggests that thi…