Summary of Power-softmax: Towards Secure Llm Inference Over Encrypted Data, by Itamar Zimerman et al.
Power-Softmax: Towards Secure LLM Inference over Encrypted Databy Itamar Zimerman, Allon Adir, Ehud Aharoni, Matan…
Power-Softmax: Towards Secure LLM Inference over Encrypted Databy Itamar Zimerman, Allon Adir, Ehud Aharoni, Matan…
Training on Fake Labels: Mitigating Label Leakage in Split Learning via Secure Dimension Transformationby Yukun…
Mind the Gap: a Spectral Analysis of Rank Collapse and Signal Propagation in Attention Layersby…
Benign Overfitting in Single-Head Attentionby Roey Magen, Shuning Shang, Zhiwei Xu, Spencer Frei, Wei Hu,…
Upcycling Large Language Models into Mixture of Expertsby Ethan He, Abhinav Khattar, Ryan Prenger, Vijay…
Differential Transformerby Tianzhu Ye, Li Dong, Yuqing Xia, Yutao Sun, Yi Zhu, Gao Huang, Furu…
EXAQ: Exponent Aware Quantization For LLMs Accelerationby Moran Shkolnik, Maxim Fishman, Brian Chmiel, Hilla Ben-Yaacov,…
On Expert Estimation in Hierarchical Mixture of Experts: Beyond Softmax Gating Functionsby Huy Nguyen, Xing…
Unveil Benign Overfitting for Transformer in Vision: Training Dynamics, Convergence, and Generalizationby Jiarui Jiang, Wei…
Cottention: Linear Transformers With Cosine Attentionby Gabriel Mongaras, Trevor Dohm, Eric C. LarsonFirst submitted to…