Mixture-of-Experts (MoE) has successfully scaled up models while maintaining nearly constant computing costs.Efficiency of MoE is challenging to achieve due to imbalanced expert activation and communication overhead.The paper proposes a collaboration-constrained routing (C2R) strategy to improve expert utilization and reduce communication costs.Experiments show an average performance improvement of 0.51% and 0.33% on two MoE models across ten NLP benchmarks.