AxMoE: Characterizing the Impact of Approximate Multipliers on Mixture-of-Experts DNN Architectures
arXiv:2605.04754v1 Announce Type: new
Abstract: Deep neural network (DNN) inference at the edge demands simultaneous improvements in accuracy, computational efficiency, and energy consumption. Approximate computing and Mixture-of-Experts (MoE) archite…