Summary of Dager: Exact Gradient Inversion For Large Language Models, by Ivo Petrov and Dimitar I. Dimitrov et al.
DAGER: Exact Gradient Inversion for Large Language Modelsby Ivo Petrov, Dimitar I. Dimitrov, Maximilian Baader,…
DAGER: Exact Gradient Inversion for Large Language Modelsby Ivo Petrov, Dimitar I. Dimitrov, Maximilian Baader,…
Linking In-context Learning in Transformers to Human Episodic Memoryby Li Ji-An, Corey Y. Zhou, Marcus…
LoRA-Ensemble: Efficient Uncertainty Modelling for Self-attention Networksby Michelle Halbheer, Dominik J. Mühlematter, Alexander Becker, Dominik…
RadarOcc: Robust 3D Occupancy Prediction with 4D Imaging Radarby Fangqiang Ding, Xiangyu Wen, Yunzhou Zhu,…
DCT-Based Decorrelated Attention for Vision Transformersby Hongyi Pan, Emadeldeen Hamdan, Xin Zhu, Koushik Biswas, Ahmet…
Next-token prediction capacity: general upper bounds and a lower bound for transformersby Liam Madden, Curtis…
Unlocking the Power of Patch: Patch-Based MLP for Long-Term Time Series Forecastingby Peiwang Tang, Weitai…
A Robust Autoencoder Ensemble-Based Approach for Anomaly Detection in Textby Jeremie Pantin, Christophe MarsalaFirst submitted…
Transformer in Touch: A Surveyby Jing Gao, Ning Cheng, Bin Fang, Wenjuan HanFirst submitted to…
Comparing Neighbors Together Makes it Easy: Jointly Comparing Multiple Candidates for Efficient and Effective Retrievalby…