Summary of Investigating Sparsity in Recurrent Neural Networks, by Harshil Darji
Investigating Sparsity in Recurrent Neural Networksby Harshil DarjiFirst submitted to arxiv on: 30 Jul 2024CategoriesMain:…
Investigating Sparsity in Recurrent Neural Networksby Harshil DarjiFirst submitted to arxiv on: 30 Jul 2024CategoriesMain:…
Realizing Unaligned Block-wise Pruning for DNN Acceleration on Mobile Devicesby Hayun Lee, Dongkun ShinFirst submitted…
SFPrompt: Communication-Efficient Split Federated Fine-Tuning for Large Pre-Trained Models over Resource-Limited Devicesby Linxiao Cao, Yifei…
A deeper look at depth pruning of LLMsby Shoaib Ahmed Siddiqui, Xin Dong, Greg Heinrich,…
Comprehensive Study on Performance Evaluation and Optimization of Model Compression: Bridging Traditional Deep Learning and…
Shapley Pruning for Neural Network Compressionby Kamil Adamczewski, Yawei Li, Luc van GoolFirst submitted to…
Compact Language Models via Pruning and Knowledge Distillationby Saurav Muralidharan, Sharath Turuvekere Sreenivas, Raviraj Joshi,…
LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inferenceby Qichen Fu, Minsik Cho, Thomas…
Reconstruct the Pruned Model without Any Retrainingby Pingjie Wang, Ziqing Fan, Shengchao Hu, Zhe Chen,…
Hybrid Dynamic Pruning: A Pathway to Efficient Transformer Inferenceby Ghadeer Jaradat, Mohammed Tolba, Ghada Alsuhli,…