Summary of Skim: Any-bit Quantization Pushing the Limits Of Post-training Quantization, by Runsheng Bai et al.
SKIM: Any-bit Quantization Pushing The Limits of Post-Training Quantizationby Runsheng Bai, Bo Liu, Qiang LiuFirst…
SKIM: Any-bit Quantization Pushing The Limits of Post-Training Quantizationby Runsheng Bai, Bo Liu, Qiang LiuFirst…
Quantized and Interpretable Learning Scheme for Deep Neural Networks in Classification Taskby Alireza Maleki, Mahsa…
CPTQuant - A Novel Mixed Precision Post-Training Quantization Techniques for Large Language Modelsby Amitash Nanda,…
Unifying KV Cache Compression for Large Language Models with LeanKVby Yanqi Zhang, Yuwei Hu, Runyuan…
FlashAttention on a Napkin: A Diagrammatic Approach to Deep Learning IO-Awarenessby Vincent Abbott, Gioele ZardiniFirst…
CEGI: Measuring the trade-off between efficiency and carbon emissions for SLMs and VLMsby Abhas Kumar,…
RILQ: Rank-Insensitive LoRA-based Quantization Error Compensation for Boosting 2-bit Large Language Model Accuracyby Geonho Lee,…
A Wave is Worth 100 Words: Investigating Cross-Domain Transferability in Time Seriesby Xiangkai Ma, Xiaobin…
DFRot: Achieving Outlier-Free and Massive Activation-Free for Rotated LLMs with Refined Rotationby Jingyang Xiang, Sai…
Training Multi-Layer Binary Neural Networks With Local Binary Error Signalsby Luca Colombo, Fabrizio Pittorino, Manuel…