LightMoE: Reducing Mixture-of-Experts Redundancy through Expert Replacing
arXiv:2603.12645v1 Announce Type: new Abstract: Mixture-of-Experts (MoE) based Large Language Models (LLMs) have demonstrated impressive performance and computational efficiency. However, their deployment is often constrained …
Jiawei Hao, Zhiwei Hao, Jianyuan Guo, Li Shen, Yong Luo, Han Hu, Dan Zeng
8 views