Summary of Towards Interpreting Language Models: a Case Study in Multi-hop Reasoning, by Mansi Sakarvadia
Towards Interpreting Language Models: A Case Study in Multi-Hop Reasoningby Mansi SakarvadiaFirst submitted to arxiv…
Towards Interpreting Language Models: A Case Study in Multi-Hop Reasoningby Mansi SakarvadiaFirst submitted to arxiv…
BitNet a4.8: 4-bit Activations for 1-bit LLMsby Hongyu Wang, Shuming Ma, Furu WeiFirst submitted to…
Clustering in Causal Attention Maskingby Nikita Karagodin, Yury Polyanskiy, Philippe RigolletFirst submitted to arxiv on:…
EffiCANet: Efficient Time Series Forecasting with Convolutional Attentionby Xinxing Zhou, Jiaqi Ye, Shubao Zhao, Ming…
Exploring Hierarchical Molecular Graph Representation in Multimodal LLMsby Chengxin Hu, Hao Li, Yihe Yuan, Jing…
Pruning Literals for Highly Efficient Explainability at Word Levelby Rohan Kumar Yadav, Bimal Bhattarai, Abhik…
Can Custom Models Learn In-Context? An Exploration of Hybrid Architecture Performance on In-Context Learning Tasksby…
How Transformers Solve Propositional Logic Problems: A Mechanistic Analysisby Guan Zhe Hong, Nishanth Dikkala, Enming…
Generalized Trusted Multi-view Classification Framework with Hierarchical Opinion Aggregationby Long Shi, Chuanqing Tang, Huangyi Deng,…
LASER: Attention with Exponential Transformationby Sai Surya Duvvuri, Inderjit S. DhillonFirst submitted to arxiv on:…