Summary of Exploring Accuracy-fairness Trade-off in Large Language Models, by Qingquan Zhang et al.
Exploring Accuracy-Fairness Trade-off in Large Language Modelsby Qingquan Zhang, Qiqi Duan, Bo Yuan, Yuhui Shi,…
Exploring Accuracy-Fairness Trade-off in Large Language Modelsby Qingquan Zhang, Qiqi Duan, Bo Yuan, Yuhui Shi,…
Multiset Transformer: Advancing Representation Learning in Persistence Diagramsby Minghua Wang, Ziyun Huang, Jinhui XuFirst submitted…
Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Modelsby Javier Ferrando, Oscar…
Hymba: A Hybrid-head Architecture for Small Language Modelsby Xin Dong, Yonggan Fu, Shizhe Diao, Wonmin…
Transformers with Sparse Attention for Granger Causalityby Riya Mahesh, Rahul Vashisht, Chandrashekar LakshminarayananFirst submitted to…
LLMSteer: Improving Long-Context LLM Inference by Steering Attention on Reused Contextsby Zhuohan Gu, Jiayi Yao,…
Selective Attention: Enhancing Transformer through Principled Context Controlby Xuechen Zhang, Xiangyu Chang, Mingchen Li, Amit…
Transformer Neural Processes - Kernel Regressionby Daniel Jenson, Jhonathan Navott, Mengyan Zhang, Makkunda Sharma, Elizaveta…
Higher Order Graph Attention Probabilistic Walk Networksby Thomas Bailie, Yun Sing Koh, Karthik MukkavilliFirst submitted…
Mechanism and Emergence of Stacked Attention Heads in Multi-Layer Transformersby Tiberiu MusatFirst submitted to arxiv…