Skip to content

ljcleo/moe-lrc

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models

This is the codebase for:

Not All Models Suit Expert Offloading:
On Local Routing Consistency of Mixture-of-Expert Models

[📄Paper] • [💻Code]

Sample routing score of GRIN-MoE and Jamba-Mini-1.6

Requirements

Setup a virtual environment with Python 3.13, and run pip install -e requirements.txt to install dependencies. You will also need scattermoe and smoe to run LLaMA-MoE-v2.

Usage

  1. Download raw data files
  2. Download model files
  3. Run scripts and notebooks!

Cite

@misc{liang2025modelssuitexpertoffloading,
      title={Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models}, 
      author={Jingcong Liang and Siyuan Wang and Miren Tian and Yitong Li and Duyu Tang and Zhongyu Wei},
      year={2025},
      eprint={2505.16056},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2505.16056}, 
}

About

Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models

Resources

License

Stars

Watchers

Forks