Summary of Explorations Of Self-repair in Language Models, by Cody Rushing et al.
Explorations of Self-Repair in Language Modelsby Cody Rushing, Neel NandaFirst submitted to arxiv on: 23…
Explorations of Self-Repair in Language Modelsby Cody Rushing, Neel NandaFirst submitted to arxiv on: 23…
TransFlower: An Explainable Transformer-Based Model with Flow-to-Flow Attention for Commuting Flow Predictionby Yan Luo, Zhuoyue…
Smoothed Graph Contrastive Learning via Seamless Proximity Integrationby Maysam Behmanesh, Maks OvsjanikovFirst submitted to arxiv…
Let’s Rectify Step by Step: Improving Aspect-based Sentiment Analysis with Diffusion Modelsby Shunyu Liu, Jie…
Attention-Guided Masked Autoencoders For Learning Image Representationsby Leon Sick, Dominik Engel, Pedro Hermosilla, Timo RopinskiFirst…
Interpreting Context Look-ups in Transformers: Investigating Attention-MLP Interactionsby Clement Neo, Shay B. Cohen, Fazl BarezFirst…
Multimodal Transformer With a Low-Computational-Cost Guaranteeby Sungjin Park, Edward ChoiFirst submitted to arxiv on: 23…
Stop Reasoning! When Multimodal LLM with Chain-of-Thought Reasoning Meets Adversarial Imageby Zefeng Wang, Zhen Han,…
MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Casesby Zechun Liu, Changsheng Zhao, Forrest…
Boosting gets full Attention for Relational Learningby Mathieu Guillame-Bert, Richard NockFirst submitted to arxiv on:…