Summary of Limit Theorems For Stochastic Gradient Descent with Infinite Variance, by Jose Blanchet et al.
Limit Theorems for Stochastic Gradient Descent with Infinite Varianceby Jose Blanchet, Aleksandar Mijatović, Wenhao YangFirst…
Limit Theorems for Stochastic Gradient Descent with Infinite Varianceby Jose Blanchet, Aleksandar Mijatović, Wenhao YangFirst…
Implicit Regularization of Sharpness-Aware Minimization for Scale-Invariant Problemsby Bingcong Li, Liang Zhang, Niao HeFirst submitted…
Double-Bayesian Learningby Stefan JaegerFirst submitted to arxiv on: 16 Oct 2024CategoriesMain: Machine Learning (cs.LG)Secondary: Neural…
Age-of-Gradient Updates for Federated Learning over Random Access Channelsby Yu Heng Wu, Houman Asgari, Stefano…
Hessian-Informed Flow Matchingby Christopher Iliffe Sprague, Arne Elofsson, Hossein AzizpourFirst submitted to arxiv on: 15…
Non-convergence to global minimizers in data driven supervised deep learning: Adam and stochastic gradient descent…
Sharpness-Aware Minimization Efficiently Selects Flatter Minima Late in Trainingby Zhanpeng Zhou, Mingze Wang, Yuchen Mao,…
Simultaneous Computation and Memory Efficient Zeroth-Order Optimizer for Fine-Tuning Large Language Modelsby Fei Wang, Li…
Learning Orthogonal Multi-Index Models: A Fine-Grained Information Exponent Analysisby Yunwei Ren, Jason D. LeeFirst submitted…
Zeroth-Order Fine-Tuning of LLMs in Random Subspacesby Ziming Yu, Pan Zhou, Sike Wang, Jia Li,…