Summary of Ttaq: Towards Stable Post-training Quantization in Continuous Domain Adaptation, by Junrui Xiao et al.
TTAQ: Towards Stable Post-training Quantization in Continuous Domain Adaptationby Junrui Xiao, Zhikai Li, Lianwei Yang,…
TTAQ: Towards Stable Post-training Quantization in Continuous Domain Adaptationby Junrui Xiao, Zhikai Li, Lianwei Yang,…
DQA: An Efficient Method for Deep Quantization of Deep Neural Network Activationsby Wenhao Hu, Paul…
CRVQ: Channel-Relaxed Vector Quantization for Extreme Compression of LLMsby Yuzhuang Xu, Shiyu Ji, Qingfu Zhu,…
Optimising TinyML with Quantization and Distillation of Transformer and Mamba Models for Indoor Localisation on…
Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionariesby Junhyuck Kim, Jongho Park,…
TurboAttention: Efficient Attention Approximation For High Throughputs LLMsby Hao Kang, Srikant Bharadwaj, James Hensman, Tushar…
Low-Rank Correction for Quantized LLMsby Meyer Scetbon, James HensmanFirst submitted to arxiv on: 10 Dec…
Post-Training Non-Uniform Quantization for Convolutional Neural Networksby Ahmed Luqman, Khuzemah Qazi, Imdadullah KhanFirst submitted to…
FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantizationby Boyang Zhang, Daning Cheng, Yunquan Zhang, Fangmin…
Taming Sensitive Weights : Noise Perturbation Fine-tuning for Robust LLM Quantizationby Dongwei Wang, Huanrui YangFirst…