Poster "mixture-of-experts models" Papers
7 papers found
DeRS: Towards Extremely Efficient Upcycled Mixture-of-Experts Models
Yongqi Huang, Peng Ye, Chenyu Huang et al.
CVPR 2025posterarXiv:2503.01359
6
citations
Mixture Compressor for Mixture-of-Experts LLMs Gains More
Wei Huang, Yue Liao, Jianhui Liu et al.
ICLR 2025posterarXiv:2410.06270
22
citations
MoEMeta: Mixture-of-Experts Meta Learning for Few-Shot Relational Learning
Han Wu, Jie Yin
NeurIPS 2025posterarXiv:2510.23013
ReMoE: Fully Differentiable Mixture-of-Experts with ReLU Routing
Ziteng Wang, Jun Zhu, Jianfei Chen
ICLR 2025posterarXiv:2412.14711
28
citations
The Labyrinth of Links: Navigating the Associative Maze of Multi-modal LLMs
HONG LI, Nanxi Li, Yuanjie Chen et al.
ICLR 2025posterarXiv:2410.01417
3
citations
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Fuzhao Xue, Zian Zheng, Yao Fu et al.
ICML 2024poster
Scaling Beyond the GPU Memory Limit for Large Mixture-of-Experts Model Training
Yechan Kim, Hwijoon Lim, Dongsu Han
ICML 2024poster