Summary of Theory on Mixture-of-experts in Continual Learning, by Hongbo Li et al.
Theory on Mixture-of-Experts in Continual Learningby Hongbo Li, Sen Lin, Lingjie Duan, Yingbin Liang, Ness…
Theory on Mixture-of-Experts in Continual Learningby Hongbo Li, Sen Lin, Lingjie Duan, Yingbin Liang, Ness…
Interpretable Cascading Mixture-of-Experts for Urban Traffic Congestion Predictionby Wenzhao Jiang, Jindong Han, Hao Liu, Tao…
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Expertsby Haoxiang Wang, Wei Xiong, Tengyang Xie, Han…
GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theoryby Haoze Wu, Zihan Qiu, Zili…
Not Eliminate but Aggregate: Post-Hoc Control over Mixture-of-Experts to Address Shortcut Shifts in Natural Language…
Graph Knowledge Distillation to Mixture of Expertsby Pavel Rumiantsev, Mark CoatesFirst submitted to arxiv on:…
Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusionby Anke Tang, Li…
Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parametersby Yixin Song, Haotong Xie, Zhengyan…
Node-wise Filtering in Graph Neural Networks: A Mixture of Experts Approachby Haoyu Han, Juanhui Li,…
Continual Traffic Forecasting via Mixture of Expertsby Sanghyun Lee, Chanyoung ParkFirst submitted to arxiv on:…