Summary of Shortened Llama: Depth Pruning For Large Language Models with Comparison Of Retraining Methods, by Bo-kyeong Kim et al.
Shortened LLaMA: Depth Pruning for Large Language Models with Comparison of Retraining Methodsby Bo-Kyeong Kim,…
Shortened LLaMA: Depth Pruning for Large Language Models with Comparison of Retraining Methodsby Bo-Kyeong Kim,…
Training-time Neuron Alignment through Permutation Subspace for Improving Linear Mode Connectivity and Model Fusionby Zexi…
Spatial-temporal-demand clustering for solving large-scale vehicle routing problems with time windowsby Christoph Kerscher, Stefan MinnerFirst…
EPSD: Early Pruning with Self-Distillation for Efficient Model Compressionby Dong Chen, Ning Liu, Yichen Zhu,…
MAPPING: Debiasing Graph Neural Networks for Fair Node Classification with Limited Sensitive Information Leakageby Ying…
Robustness to distribution shifts of compressed networks for edge devicesby Lulan Shen, Ali Edalati, Brett…
APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inferenceby Bowen Zhao,…
Zero-Space Cost Fault Tolerance for Transformer-based Language Models on ReRAMby Bingbing Li, Geng Yuan, Zigeng…
Pruning for Protection: Increasing Jailbreak Resistance in Aligned LLMs Without Fine-Tuningby Adib Hasan, Ileana Rugina,…
SymbolNet: Neural Symbolic Regression with Adaptive Dynamic Pruning for Compressionby Ho Fung Tsoi, Vladimir Loncar,…