Loading Now

Summary of Clarify: Improving Model Robustness with Natural Language Corrections, by Yoonho Lee et al.


Clarify: Improving Model Robustness With Natural Language Corrections

by Yoonho Lee, Michelle S. Lam, Helena Vasconcelos, Michael S. Bernstein, Chelsea Finn

First submitted to arxiv on: 6 Feb 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Artificial Intelligence (cs.AI); Computation and Language (cs.CL)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
In this paper, researchers propose Clarify, a new approach to teaching machine learning models that leverages human feedback at the concept level. Unlike previous methods that require manual labeling of training data, Clarify allows users to provide short text descriptions of model misconceptions, which are then used to improve the training process. The authors demonstrate the effectiveness of Clarify in two datasets and conduct a case study on ImageNet, finding and rectifying 31 novel hard subpopulations.
Low GrooveSquid.com (original content) Low Difficulty Summary
Machine learning models can learn incorrect ideas if they’re taught with misleading data. To fix this, we need to give them extra information. One way is by adding labels for things that are wrong or providing corrected data. But this takes a lot of work. The researchers think people can do better than that. They propose an easy way to correct model mistakes using text descriptions. It’s called Clarify. Users just write a short sentence about what the model gets wrong, and then the computer fixes it automatically. This is the first system that lets users correct models in one step. Studies show that regular people can use Clarify to make models better.

Keywords

* Artificial intelligence  * Machine learning