Loading Now

Summary of Multi-level Explanations For Generative Language Models, by Lucas Monteiro Paes et al.


Multi-Level Explanations for Generative Language Models

by Lucas Monteiro Paes, Dennis Wei, Hyo Jin Do, Hendrik Strobelt, Ronny Luss, Amit Dhurandhar, Manish Nagireddy, Karthikeyan Natesan Ramamurthy, Prasanna Sattigeri, Werner Geyer, Soumya Ghosh

First submitted to arxiv on: 21 Mar 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This work extends perturbation-based explanation methods, such as LIME and SHAP, to generative language models. To address the challenges of text output and long input texts, a general framework called MExGen is proposed, which can be instantiated with different attribution algorithms. The framework introduces scalarizers for mapping text to real numbers to handle text output and takes a multi-level approach to handle long inputs. A systematic evaluation of perturbation-based attribution methods is conducted for summarization and context-grounded question answering tasks. The results show that MExGen can provide more locally faithful explanations of generated outputs.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper helps us understand how artificial intelligence models, like language generators, make predictions. Right now, we don’t fully know why these models generate certain texts or answers. This research proposes a new way to explain the decisions made by these models. It’s like trying to figure out why someone wrote a specific sentence or answered a question in a particular way. The method is designed to work with different types of language generation tasks and can help us better understand how these models make predictions.

Keywords

» Artificial intelligence  » Question answering  » Summarization