Loading Now

Summary of The Effect Of Model Size on Llm Post-hoc Explainability Via Lime, by Henning Heyen et al.


The Effect of Model Size on LLM Post-hoc Explainability via LIME

by Henning Heyen, Amy Widdicombe, Noah Y. Siegel, Maria Perez-Ortiz, Philip Treleaven

First submitted to arxiv on: 8 May 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The abstract discusses how large language models (LLMs) become more performant with increased size, but there’s a gap in understanding how explainability is affected by this trend. Specifically, the authors investigate LIME explanations for DeBERTaV3 models of varying sizes on natural language inference (NLI) and zero-shot classification (ZSC) tasks. They evaluate these explanations based on faithfulness to internal decision processes and plausibility, aligning with human explanations. The key finding is that larger models do not necessarily lead to more plausible explanations despite improved performance, indicating a disconnection between LIME explanations and the models’ internal workings as size increases. Furthermore, the study highlights limitations in NLI contexts regarding faithfulness metrics.
Low GrooveSquid.com (original content) Low Difficulty Summary
Large language models are getting bigger, but what happens to their ability to explain themselves? This paper looks at how well smaller and larger DeBERTaV3 models can be explained using LIME (Local Interpretable Model-agnostic Explanations) on two tasks: understanding sentences and predicting what they mean. The researchers want to know if these explanations are faithful to the model’s own thinking and make sense to humans. They found that bigger models don’t necessarily create more believable explanations, even though they do better at the task. This raises questions about how well we can understand how these large language models work.

Keywords

» Artificial intelligence  » Classification  » Inference  » Zero shot