cs.LG

PreMoE: Proactive Inference for Efficient Mixture-of-Experts

arXiv:2505.17639v3 Announce Type: replace
Abstract: Mixture-of-Experts (MoE) models offer dynamic computation, but are typically deployed as static full-capacity models, missing opportunities for deployment-specific specialization. We introduce PreMoE…

Scroll to Top