Summary of Dager: Exact Gradient Inversion For Large Language Models, by Ivo Petrov and Dimitar I. Dimitrov et al.
DAGER: Exact Gradient Inversion for Large Language Modelsby Ivo Petrov, Dimitar I. Dimitrov, Maximilian Baader,…
DAGER: Exact Gradient Inversion for Large Language Modelsby Ivo Petrov, Dimitar I. Dimitrov, Maximilian Baader,…
Towards Understanding the Working Mechanism of Text-to-Image Diffusion Modelby Mingyang Yi, Aoxue Li, Yi Xin,…
iVideoGPT: Interactive VideoGPTs are Scalable World Modelsby Jialong Wu, Shaofeng Yin, Ningya Feng, Xu He,…
MultiCast: Zero-Shot Multivariate Time Series Forecasting Using LLMsby Georgios Chatzigeorgakidis, Konstantinos Lentzos, Dimitrios SkoutasFirst submitted…
Unchosen Experts Can Contribute Too: Unleashing MoE Models’ Power by Self-Contrastby Chufan Shi, Cheng Yang,…
Segformer++: Efficient Token-Merging Strategies for High-Resolution Semantic Segmentationby Daniel Kienzle, Marco Kantonis, Robin Schön, Rainer…
Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Modelsby Yongxin Guo, Zhenglin Cheng,…
Self-Taught Recognizer: Toward Unsupervised Adaptation for Speech Foundation Modelsby Yuchen Hu, Chen Chen, Chao-Han Huck…
Next-token prediction capacity: general upper bounds and a lower bound for transformersby Liam Madden, Curtis…
Asymptotic theory of in-context learning by linear attentionby Yue M. Lu, Mary I. Letey, Jacob…