Summary of Zeroth-order Fine-tuning Of Llms with Extreme Sparsity, by Wentao Guo et al.
Zeroth-Order Fine-Tuning of LLMs with Extreme Sparsityby Wentao Guo, Jikai Long, Yimeng Zeng, Zirui Liu,…
Zeroth-Order Fine-Tuning of LLMs with Extreme Sparsityby Wentao Guo, Jikai Long, Yimeng Zeng, Zirui Liu,…
SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretrainingby Andi Han, Jiaxiang…
MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantizationby Aozhong Zhang, Naigang Wang, Yanxia Deng, Xin…
Outliers and Calibration Sets have Diminishing Effect on Quantization of Modern LLMsby Davide Paglieri, Saurabh…
Effective Interplay between Sparsity and Quantization: From Theory to Practiceby Simla Burcu Harma, Ayan Chakraborty,…
LCQ: Low-Rank Codebook based Quantization for Large Language Modelsby Wen-Pu Cai, Ming-Yang Li, Wu-Jun LiFirst…
Understanding and Minimising Outlier Features in Neural Network Trainingby Bobby He, Lorenzo Noci, Daniele Paliotta,…
4-bit Shampoo for Memory-Efficient Network Trainingby Sike Wang, Pan Zhou, Jia Li, Hua HuangFirst submitted…
Exploiting LLM Quantizationby Kazuki Egashira, Mark Vero, Robin Staab, Jingxuan He, Martin VechevFirst submitted to…
I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit Large Language Modelsby Xing Hu, Yuan Cheng, Dawei…