ReGATE: Learning Faster and Better with Fewer Tokens in MLLMs
arXiv:2507.21420v3 Announce Type: replace-cross
Abstract: The computational cost of training multimodal large language models (MLLMs) grows rapidly with the number of processed tokens. Existing efficiency methods mainly target inference via token redu…