Summary of Fusemoe: Mixture-of-experts Transformers For Fleximodal Fusion, by Xing Han et al.
FuseMoE: Mixture-of-Experts Transformers for Fleximodal Fusionby Xing Han, Huy Nguyen, Carl Harris, Nhat Ho, Suchi…
FuseMoE: Mixture-of-Experts Transformers for Fleximodal Fusionby Xing Han, Huy Nguyen, Carl Harris, Nhat Ho, Suchi…
InterpretCC: Intrinsic User-Centric Interpretability through Global Mixture of Expertsby Vinitra Swamy, Syrielle Montariol, Julian Blackwell,…
On Least Square Estimation in Softmax Gating Mixture of Expertsby Huy Nguyen, Nhat Ho, Alessandro…
CompeteSMoE – Effective Training of Sparse Mixture of Experts via Competitionby Quang Pham, Giang Do,…
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Modelsby Fuzhao Xue, Zian Zheng, Yao Fu,…
MoDE: A Mixture-of-Experts Model with Mutual Distillation among the Expertsby Zhitian Xie, Yinger Zhang, Chenyi…
Merging Multi-Task Models via Weight-Ensembling Mixture of Expertsby Anke Tang, Li Shen, Yong Luo, Nan…
Explainable data-driven modeling via mixture of experts: towards effective blending of grey and black-box modelsby…
Checkmating One, by Using Many: Combining Mixture of Experts with MCTS to Improve in Chessby…
Routers in Vision Mixture of Experts: An Empirical Studyby Tianlin Liu, Mathieu Blondel, Carlos Riquelme,…