Loading Now

Summary of The Group Robustness Is in the Details: Revisiting Finetuning Under Spurious Correlations, by Tyler Labonte et al.


The Group Robustness is in the Details: Revisiting Finetuning under Spurious Correlations

by Tyler LaBonte, John C. Hill, Xinchen Zhang, Vidya Muthukumar, Abhishek Kumar

First submitted to arxiv on: 19 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper investigates the performance of finetuned machine learning models on minority groups, finding that they often rely too heavily on spurious correlations. To address this issue, the authors conduct comprehensive experiments across four benchmarks in vision and language tasks. They first show that commonly used class-balancing techniques can actually decrease worst-group accuracy (WGA) with training epochs, suggesting that these approaches may not be effective. The authors then propose a mixture method that can outperform both techniques in certain scenarios. Additionally, they find that scaling pre-trained models is generally beneficial for WGA, but only when combined with appropriate class-balancing. The paper also identifies spectral imbalance in finetuning features as a potential source of group disparities. Overall, the results highlight more nuanced interactions between modern finetuned models and group robustness than previously understood.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper looks at how well machine learning models work for different groups of people. Right now, many models are good at recognizing patterns that aren’t actually important, which can lead to poor performance on certain groups. The authors did lots of experiments on four big datasets in vision and language tasks to see what happens when they fine-tune their models. They found that some common ways to balance the data don’t work well, and that scaling pre-trained models can help, but only if you do it the right way. The results show that these modern models are more complex than we thought, and that there’s still a lot to learn about making them fairer.

Keywords

» Artificial intelligence  » Machine learning