Summary of On the Learn-to-optimize Capabilities Of Transformers in In-context Sparse Recovery, by Renpu Liu et al.
On the Learn-to-Optimize Capabilities of Transformers in In-Context Sparse Recoveryby Renpu Liu, Ruida Zhou, Cong…
On the Learn-to-Optimize Capabilities of Transformers in In-Context Sparse Recoveryby Renpu Liu, Ruida Zhou, Cong…
Low-Rank Adversarial PGD Attackby Dayana Savostianova, Emanuele Zangrando, Francesco TudiscoFirst submitted to arxiv on: 16…
State-space models can learn in-context by gradient descentby Neeraj Mohan Sushma, Yudou Tian, Harshvardhan Mestha,…
On the Training Convergence of Transformers for In-Context Classification of Gaussian Mixturesby Wei Shen, Ruida…
Beyond Linear Approximations: A Novel Pruning Approach for Attention Matrixby Yingyu Liang, Jiangxuan Long, Zhenmei…
Bypassing the Exponential Dependency: Looped Transformers Efficiently Learn In-context by Multi-step Gradient Descentby Bo Chen,…
Fast Second-Order Online Kernel Learning through Incremental Matrix Sketching and Decompositionby Dongxie Wen, Xiao Zhang,…
ROSAR: An Adversarial Re-Training Framework for Robust Side-Scan Sonar Object Detectionby Martin Aubard, László Antal,…
Robust Gradient Descent for Phase Retrievalby Alex Buna, Patrick RebeschiniFirst submitted to arxiv on: 14…
Feature Averaging: An Implicit Bias of Gradient Descent Leading to Non-Robustness in Neural Networksby Binghui…