Summary of Two Are Better Than One: Context Window Extension with Multi-grained Self-injection, by Wei Han et al.
Two are better than one: Context window extension with multi-grained self-injectionby Wei Han, Pan Zhou,…
Two are better than one: Context window extension with multi-grained self-injectionby Wei Han, Pan Zhou,…
TesseraQ: Ultra Low-Bit LLM Post-Training Quantization with Block Reconstructionby Yuhang Li, Priyadarshini PandaFirst submitted to…
Knowledge Distillation Using Frontier Open-source LLMs: Generalizability and the Role of Synthetic Databy Anup Shirgaonkar,…
Prompting and Fine-Tuning of Small LLMs for Length-Controllable Telephone Call Summarizationby David Thulke, Yingbo Gao,…
Beware of Calibration Data for Pruning Large Language Modelsby Yixin Ji, Yang Xiang, Juntao Li,…
Representation Shattering in Transformers: A Synthetic Study with Knowledge Editingby Kento Nishi, Maya Okawa, Rahul…
FastAttention: Extend FlashAttention2 to NPUs and Low-resource GPUsby Haoran Lin, Xianzhi Yu, Kang Zhao, Lu…
Large Language Models in Computer Science Education: A Systematic Literature Reviewby Nishat Raihan, Mohammed Latif…
Natural GaLore: Accelerating GaLore for memory-efficient LLM Training and Fine-tuningby Arijit DasFirst submitted to arxiv…
Transit Pulse: Utilizing Social Media as a Source for Customer Feedback and Information Extraction with…