menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Advancing ...
source image

Arxiv

1d

read

262

img
dot

Image Credit: Arxiv

Advancing MoE Efficiency: A Collaboration-Constrained Routing (C2R) Strategy for Better Expert Parallelism Design

  • Mixture-of-Experts (MoE) has successfully scaled up models while maintaining nearly constant computing costs.
  • Efficiency of MoE is challenging to achieve due to imbalanced expert activation and communication overhead.
  • The paper proposes a collaboration-constrained routing (C2R) strategy to improve expert utilization and reduce communication costs.
  • Experiments show an average performance improvement of 0.51% and 0.33% on two MoE models across ten NLP benchmarks.

Read Full Article

like

15 Likes

For uninterrupted reading, download the app