HI-MoE: Hierarchical Instance-Conditioned Mixture-of-Experts for Object Detection
arXiv:2604.04908v1 Announce Type: new
Abstract: Mixture-of-Experts (MoE) architectures enable conditional computation by activating only a subset of model parameters for each input. Although sparse routing has been highly effective in language models …