Summary of On the Compressibility Of Quantized Large Language Models, by Yu Mao et al.
On the Compressibility of Quantized Large Language Modelsby Yu Mao, Weilan Wang, Hongchao Du, Nan…
On the Compressibility of Quantized Large Language Modelsby Yu Mao, Weilan Wang, Hongchao Du, Nan…
LLM-PQ: Serving LLM on Heterogeneous Clusters with Phase-Aware Partition and Adaptive Quantizationby Juntao Zhao, Borui…
Inexact Unlearning Needs More Careful Evaluations to Avoid a False Sense of Privacyby Jamie Hayes,…
Defending Against Data Reconstruction Attacks in Federated Learning: An Information Theory Approachby Qi Tan, Qi…
MALTO at SemEval-2024 Task 6: Leveraging Synthetic Data for LLM Hallucination Detectionby Federico Borra, Claudio…
Teaching MLP More Graph Information: A Three-stage Multitask Knowledge Distillation Frameworkby Junxian Li, Bin Shi,…
Pairwise Alignment Improves Graph Domain Adaptationby Shikun Liu, Deyu Zou, Han Zhao, Pan LiFirst submitted…
Direct Alignment of Draft Model for Speculative Decoding with Chat-Fine-Tuned LLMsby Raghavv Goel, Mukul Gagrani,…
Resilience of Entropy Model in Distributed Neural Networksby Milin Zhang, Mohammad Abdi, Shahriar Rifat, Francesco…
Stability-Certified Learning of Control Systems with Quadratic Nonlinearitiesby Igor Pontes Duff, Pawan Goyal, Peter BennerFirst…