Summary of Shadowllm: Predictor-based Contextual Sparsity For Large Language Models, by Yash Akhauri et al.
ShadowLLM: Predictor-based Contextual Sparsity for Large Language Modelsby Yash Akhauri, Ahmed F AbouElhamayed, Jordan Dotzel,…
ShadowLLM: Predictor-based Contextual Sparsity for Large Language Modelsby Yash Akhauri, Ahmed F AbouElhamayed, Jordan Dotzel,…
AutoDetect: Towards a Unified Framework for Automated Weakness Detection in Large Language Modelsby Jiale Cheng,…
What Matters in Transformers? Not All Attention is Neededby Shwai He, Guoheng Sun, Zheyu Shen,…
LaMSUM: Amplifying Voices Against Harassment through LLM Guided Extractive Summarization of User Incident Reportsby Garima…
Steering Without Side Effects: Improving Post-Deployment Control of Language Modelsby Asa Cooper Stickland, Alexander Lyzhov,…
Optimised Grouped-Query Attention Mechanism for Transformersby Yuang Chen, Cheng Zhang, Xitong Gao, Robert D. Mullins,…
Domain Adaptation of Llama3-70B-Instruct through Continual Pre-Training and Model Merging: A Comprehensive Evaluationby Shamane Siriwardhana,…
Understanding Finetuning for Factual Knowledge Extractionby Gaurav Ghosal, Tatsunori Hashimoto, Aditi RaghunathanFirst submitted to arxiv…
Can Low-Rank Knowledge Distillation in LLMs be Useful for Microelectronic Reasoning?by Nirjhor Rouf, Fin Amin,…
MOYU: A Theoretical Study on Massive Over-activation Yielded Uplifts in LLMsby Chi Ma, Mincong Huang,…