Summary of Zeroth-order Fine-tuning Of Llms with Extreme Sparsity, by Wentao Guo et al.
Zeroth-Order Fine-Tuning of LLMs with Extreme Sparsityby Wentao Guo, Jikai Long, Yimeng Zeng, Zirui Liu,…
Zeroth-Order Fine-Tuning of LLMs with Extreme Sparsityby Wentao Guo, Jikai Long, Yimeng Zeng, Zirui Liu,…
SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretrainingby Andi Han, Jiaxiang…
MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantizationby Aozhong Zhang, Naigang Wang, Yanxia Deng, Xin…
Outliers and Calibration Sets have Diminishing Effect on Quantization of Modern LLMsby Davide Paglieri, Saurabh…
Effective Interplay between Sparsity and Quantization: From Theory to Practiceby Simla Burcu Harma, Ayan Chakraborty,…
LCQ: Low-Rank Codebook based Quantization for Large Language Modelsby Wen-Pu Cai, Ming-Yang Li, Wu-Jun LiFirst…
Understanding and Minimising Outlier Features in Neural Network Trainingby Bobby He, Lorenzo Noci, Daniele Paliotta,…
Exploiting LLM Quantizationby Kazuki Egashira, Mark Vero, Robin Staab, Jingxuan He, Martin VechevFirst submitted to…
4-bit Shampoo for Memory-Efficient Network Trainingby Sike Wang, Pan Zhou, Jia Li, Hua HuangFirst submitted…
I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit Large Language Modelsby Xing Hu, Yuan Cheng, Dawei…