Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencentby Xingwu Sun, Yanfeng…
Summary of Efficient Mixture-of-expert For Video-based Driver State and Physiological Multi-task Estimation in Conditional Autonomous Driving, by Jiyao Wang et al.
Efficient Mixture-of-Expert for Video-based Driver State and Physiological Multi-task Estimation in Conditional Autonomous Drivingby Jiyao…
Summary of Expertflow: Optimized Expert Activation and Token Allocation For Efficient Mixture-of-experts Inference, by Xin He et al.
ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inferenceby Xin He, Shunkang Zhang,…
Summary of Scalable Multi-domain Adaptation Of Language Models Using Modular Experts, by Peter Schafhalter et al.
Scalable Multi-Domain Adaptation of Language Models using Modular Expertsby Peter Schafhalter, Shun Liao, Yanqi Zhou,…
Summary of Scidfm: a Large Language Model with Mixture-of-experts For Science, by Liangtai Sun et al.
SciDFM: A Large Language Model with Mixture-of-Experts for Scienceby Liangtai Sun, Danyu Luo, Da Ma,…
Summary of Chartmoe: Mixture Of Diversely Aligned Expert Connector For Chart Understanding, by Zhengzhuo Xu et al.
ChartMoE: Mixture of Diversely Aligned Expert Connector for Chart Understandingby Zhengzhuo Xu, Bowen Qu, Yiyan…
Summary of Dutytte: Deciphering Uncertainty in Origin-destination Travel Time Estimation, by Xiaowei Mao et al.
DutyTTE: Deciphering Uncertainty in Origin-Destination Travel Time Estimationby Xiaowei Mao, Yan Lin, Shengnan Guo, Yubin…
Summary of Aquilamoe: Efficient Training For Moe Models with Scale-up and Scale-out Strategies, by Bo-wen Zhang et al.
AquilaMoE: Efficient Training for MoE Models with Scale-Up and Scale-Out Strategiesby Bo-Wen Zhang, Liangdong Wang,…
Summary of Multimodal Fusion and Coherence Modeling For Video Topic Segmentation, by Hai Yu et al.
Multimodal Fusion and Coherence Modeling for Video Topic Segmentationby Hai Yu, Chong Deng, Qinglin Zhang,…
Summary of Pmoe: Progressive Mixture Of Experts with Asymmetric Transformer For Continual Learning, by Min Jae Jung et al.
PMoE: Progressive Mixture of Experts with Asymmetric Transformer for Continual Learningby Min Jae Jung, JooHee…