Summary of Coat: Compressing Optimizer States and Activation For Memory-efficient Fp8 Training, by Haocheng Xi et al.
COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Trainingby Haocheng Xi, Han Cai, Ligeng…
COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Trainingby Haocheng Xi, Han Cai, Ligeng…
Inference time LLM alignment in single and multidomain preference spectrumby Sadat Shahriar, Zheng Qi, Nikolaos…
On the Crucial Role of Initialization for Matrix Factorizationby Bingcong Li, Liang Zhang, Aryan Mokhtari,…
From Imitation to Introspection: Probing Self-Consciousness in Language Modelsby Sirui Chen, Shu Yu, Shengjie Zhao,…
Aligning CodeLLMs with Direct Preference Optimizationby Yibo Miao, Bofei Gao, Shanghaoran Quan, Junyang Lin, Daoguang…
SAMG: Offline-to-Online Reinforcement Learning via State-Action-Conditional Offline Model Guidanceby Liyu Zhang, Haochi Wu, Xu Wan,…
3D Shape Completion with Test-Time Trainingby Michael Schopf-Kuester, Zorah Lähner, Michael MoellerFirst submitted to arxiv…
Towards Understanding the Fragility of Multilingual LLMs against Fine-Tuning Attacksby Samuele Poppi, Zheng-Xin Yong, Yifei…
LEGO: Language Model Building Blocksby Shrenik Bhansali, Alwin Jin, Tyler Lizzo, Larry HeckFirst submitted to…
Aggregated Knowledge Model: Enhancing Domain-Specific QA with Fine-Tuned and Retrieval-Augmented Generation Modelsby Fengchen Liu, Jordan…