Summary of One Step Learning, One Step Review, by Xiaolong Huang et al.
One Step Learning, One Step Reviewby Xiaolong Huang, Qiankun Li, Xueran Li, Xuesong GaoFirst submitted…
One Step Learning, One Step Reviewby Xiaolong Huang, Qiankun Li, Xueran Li, Xuesong GaoFirst submitted…
OrchMoE: Efficient Multi-Adapter Learning with Task-Skill Synergyby Haowen Wang, Tao Sun, Kaixiang Ji, Jian Wang,…
AutoFT: Learning an Objective for Robust Fine-Tuningby Caroline Choi, Yoonho Lee, Annie Chen, Allan Zhou,…
A Fast, Performant, Secure Distributed Training Framework For Large Language Modelby Wei Huang, Yinggui Wang,…
Risk-Aware Accelerated Wireless Federated Learning with Heterogeneous Clientsby Mohamed Ads, Hesham ElSawy, Hossam S. HassaneinFirst…
MADA: Meta-Adaptive Optimizers through hyper-gradient Descentby Kaan Ozkara, Can Karakus, Parameswaran Raman, Mingyi Hong, Shoham…
RAG vs Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agricultureby Angels Balaguer, Vinamra Benara,…
Contrastive Perplexity for Controlled Generation: An Application in Detoxifying Large Language Modelsby Tassilo Klein, Moin…
Activations and Gradients Compression for Model-Parallel Trainingby Mikhail Rudakov, Aleksandr Beznosikov, Yaroslav Kholodov, Alexander GasnikovFirst…
PDE Generalization of In-Context Operator Networks: A Study on 1D Scalar Nonlinear Conservation Lawsby Liu…