Summary of The Impact Of Quantization and Pruning on Deep Reinforcement Learning Models, by Heng Lu et al.
The Impact of Quantization and Pruning on Deep Reinforcement Learning Modelsby Heng Lu, Mehdi Alemi,…
The Impact of Quantization and Pruning on Deep Reinforcement Learning Modelsby Heng Lu, Mehdi Alemi,…
QET: Enhancing Quantized LLM Parameters and KV cache Compression through Element Substitution and Residual Clusteringby…
How Does Quantization Affect Multilingual LLMs?by Kelly Marchisio, Saurabh Dash, Hongyu Chen, Dennis Aumiller, Ahmet…
GPTQT: Quantize Large Language Models Twice to Push the Efficiencyby Yipin Guo, Yilin Lang, Qinyuan…
SFC: Achieve Accurate Fast Convolution under Low-precision Arithmeticby Liulu He, Yufei Zhao, Rui Gao, Yuan…
QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devicesby Juntao Zhao, Borui Wan, Yanghua Peng, Haibin…
Joint Pruning and Channel-wise Mixed-Precision Quantization for Efficient Deep Neural Networksby Beatrice Alessandra Motetti, Matteo…
LLMEasyQuant – An Easy to Use Toolkit for LLM Quantizationby Dong Liu, Kaiser PisterFirst submitted…
Reliable edge machine learning hardware for scientific applicationsby Tommaso Baldi, Javier Campos, Ben Hawks, Jennifer…
FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantizationby Linping Qu, Shenghui…