Summary of An Image Is Worth More Than 16×16 Patches: Exploring Transformers on Individual Pixels, by Duy-kien Nguyen et al.
An Image is Worth More Than 16x16 Patches: Exploring Transformers on Individual Pixelsby Duy-Kien Nguyen,…
An Image is Worth More Than 16x16 Patches: Exploring Transformers on Individual Pixelsby Duy-Kien Nguyen,…
Learning positional encodings in transformers depends on initializationby Takuya Ito, Luca Cocchi, Tim Klinger, Parikshit…
A Concept-Based Explainability Framework for Large Multimodal Modelsby Jayneel Parekh, Pegah Khayatan, Mustafa Shukor, Alasdair…
REAL Sampling: Boosting Factuality and Diversity of Open-Ended Generation via Asymptotic Entropyby Haw-Shiuan Chang, Nanyun…
Entropy-Reinforced Planning with Large Language Models for Drug Discoveryby Xuefeng Liu, Chih-chan Tien, Peng Ding,…
Transformers Provably Learn Sparse Token Selection While Fully-Connected Nets Cannotby Zixuan Wang, Stanley Wei, Daniel…
Scaling the Vocabulary of Non-autoregressive Models for Efficient Generative Retrievalby Ravisri Valluri, Akash Kumar Mohankumar,…
REP: Resource-Efficient Prompting for Rehearsal-Free Continual Learningby Sungho Jeon, Xinyue Ma, Kwang In Kim, Myeongjae…
Transformers need glasses! Information over-squashing in language tasksby Federico Barbero, Andrea Banino, Steven Kapturowski, Dharshan…
What Should Embeddings Embed? Autoregressive Models Represent Latent Generating Distributionsby Liyi Zhang, Michael Y. Li,…