Summary of Skim: Any-bit Quantization Pushing the Limits Of Post-training Quantization, by Runsheng Bai et al.
SKIM: Any-bit Quantization Pushing The Limits of Post-Training Quantizationby Runsheng Bai, Bo Liu, Qiang LiuFirst…
SKIM: Any-bit Quantization Pushing The Limits of Post-Training Quantizationby Runsheng Bai, Bo Liu, Qiang LiuFirst…
DFRot: Achieving Outlier-Free and Massive Activation-Free for Rotated LLMs with Refined Rotationby Jingyang Xiang, Sai…
COAP: Memory-Efficient Training with Correlation-Aware Gradient Projectionby Jinqi Xiao, Shen Sang, Tiancheng Zhi, Jing Liu,…
Attamba: Attending To Multi-Token Statesby Yash Akhauri, Safeen Huda, Mohamed S. AbdelfattahFirst submitted to arxiv…
Pushing the Limits of Large Language Model Quantization via the Linearity Theoremby Vladimir Malinovskii, Andrei…
CLOVER: Cross-Layer Orthogonal Vectors Pruning and Fine-Tuningby Fanxu Meng, Pingzhi Tang, Fan jiang, Muhan ZhangFirst…
Is Training Data Quality or Quantity More Impactful to Small Language Model Performance?by Aryan Sajith,…
FuseGPT: Learnable Layers Fusion of Generative Pre-trained Transformersby Zehua Pei, Hui-Ling Zhen, Xianzhi Yu, Sinno…
Deriving Activation Functions Using Integrationby Allen Hao Huang, Imanol SchlagFirst submitted to arxiv on: 20…
Efficient Adaptive Optimization via Subset-Norm and Subspace-Momentum: Fast, Memory-Reduced Training with Convergence Guaranteesby Thien Hang…