Loading Now

Summary of Self-supervised Interpretable Concept-based Models For Text Classification, by Francesco De Santis et al.


Self-supervised Interpretable Concept-based Models for Text Classification

by Francesco De Santis, Philippe Bich, Gabriele Ciravegna, Pietro Barbiero, Danilo Giordano, Tania Cerquitelli

First submitted to arxiv on: 20 Jun 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed Interpretable Concept Embedding Models (ICEMs) aim to overcome the limitations of Large-Language Models (LLMs) by providing explainable-by-design architectures for textual data. By leveraging the generalization abilities of LLMs, ICEMs predict concept labels in a self-supervised manner while delivering final predictions with an interpretable function. The models achieve similar performance to fully supervised and end-to-end black-box models, offering logical explanations for their predictions. Additionally, ICEMs are interactable, allowing humans to modify intermediate predictions through concept interventions, and controllable, guiding the LLMs’ decoding process.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper proposes a new way to make Large-Language Models (LLMs) more understandable. Instead of using traditional methods that try to explain how the model works, the authors create a new type of model called Interpretable Concept Embedding Models (ICEMs). These models can predict what’s going on inside the LLM and why it made certain decisions. This helps make the LLM more reliable and easier to use.

Keywords

» Artificial intelligence  » Embedding  » Generalization  » Self supervised  » Supervised