Summary of Beyond Autoregression: Fast Llms Via Self-distillation Through Time, by Justin Deschenaux et al.
Beyond Autoregression: Fast LLMs via Self-Distillation Through Timeby Justin Deschenaux, Caglar GulcehreFirst submitted to arxiv…
Beyond Autoregression: Fast LLMs via Self-Distillation Through Timeby Justin Deschenaux, Caglar GulcehreFirst submitted to arxiv…
Improving Multimodal Large Language Models Using Continual Learningby Shikhar Srivastava, Md Yousuf Harun, Robik Shrestha,…
Knowledge Distillation Using Frontier Open-source LLMs: Generalizability and the Role of Synthetic Databy Anup Shirgaonkar,…
Scaling up Masked Diffusion Models on Textby Shen Nie, Fengqi Zhu, Chao Du, Tianyu Pang,…
Beyond Autoregression: Discrete Diffusion for Complex Reasoning and Planningby Jiacheng Ye, Jiahui Gao, Shansan Gong,…
Enhancing Text Generation in Joint NLG/NLU Learning Through Curriculum Learning, Semi-Supervised Training, and Advanced Optimization…
H2OVL-Mississippi Vision Language Models Technical Reportby Shaikat Galib, Shanshan Wang, Guanshuo Xu, Pascal Pfeiffer, Ryan…
TextLap: Customizing Language Models for Text-to-Layout Planningby Jian Chen, Ruiyi Zhang, Yufan Zhou, Jennifer Healey,…
TapWeight: Reweighting Pretraining Objectives for Task-Adaptive Pretrainingby Ruiyi Zhang, Sai Ashish Somayajula, Pengtao XieFirst submitted…
MTL-LoRA: Low-Rank Adaptation for Multi-Task Learningby Yaming Yang, Dilxat Muhtar, Yelong Shen, Yuefeng Zhan, Jianfeng…