Summary of Inattention: Linear Context Scaling For Transformers, by Joseph Eisner
InAttention: Linear Context Scaling for Transformersby Joseph EisnerFirst submitted to arxiv on: 9 Oct 2024CategoriesMain:…
InAttention: Linear Context Scaling for Transformersby Joseph EisnerFirst submitted to arxiv on: 9 Oct 2024CategoriesMain:…
Stuffed Mamba: State Collapse and State Capacity of RNN-Based Long-Context Modelingby Yingfa Chen, Xinrong Zhang,…
Joint Fine-tuning and Conversion of Pretrained Speech and Language Models towards Linear Complexityby Mutian He,…
MatMamba: A Matryoshka State Space Modelby Abhinav Shukla, Sai Vemprala, Aditya Kusupati, Ashish KapoorFirst submitted…
Defending Membership Inference Attacks via Privacy-aware Sparsity Tuningby Qiang Hu, Hengxiang Zhang, Hongxin WeiFirst submitted…
InstantIR: Blind Image Restoration with Instant Generative Referenceby Jen-Yuan Huang, Haofan Wang, Qixun Wang, Xu…
Conformal Prediction: A Data Perspectiveby Xiaofan Zhou, Baiting Chen, Yu Gui, Lu ChengFirst submitted to…
Sketch to Adapt: Fine-Tunable Sketches for Efficient LLM Adaptationby Tianyi Zhang, Junda Su, Aditya Desai,…
EVOLvE: Evaluating and Optimizing LLMs For Explorationby Allen Nie, Yi Su, Bo Chang, Jonathan N.…
Mixture Compressor for Mixture-of-Experts LLMs Gains Moreby Wei Huang, Yue Liao, Jianhui Liu, Ruifei He,…