Rethinking Network Topologies for Cost-Effective Mixture-of-Experts LLM Serving
arXiv:2605.00254v1 Announce Type: cross
Abstract: Mixture-of-experts (MoE) architectures have turned LLM serving into a cluster-scale workload in which communication consumes a considerable portion of LLM serving runtime. This has prompted industry to…