Loading Now

Summary of Improving Llm Group Fairness on Tabular Data Via In-context Learning, by Valeriia Cherepanova et al.


Improving LLM Group Fairness on Tabular Data via In-Context Learning

by Valeriia Cherepanova, Chia-Jung Lee, Nil-Jana Akpinar, Riccardo Fogliato, Martin Andres Bertran, Michael Kearns, James Zou

First submitted to arxiv on: 5 Dec 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Large language models (LLMs) have been shown to be effective on tabular prediction tasks in low-data regimes, leveraging internal knowledge and learning from instructions. However, LLMs can fail to generate predictions that satisfy group fairness, producing inequitable outcomes across groups. Conventional debiasing approaches for natural language tasks do not directly translate to mitigating group unfairness in tabular settings. This work investigates four empirical approaches to improve group fairness of LLM predictions on tabular datasets: fair prompt optimization, soft prompt tuning, strategic selection of few-shot examples, and self-refining predictions via chain-of-thought reasoning. Experiments on four tabular datasets using open-source and proprietary LLMs show the effectiveness of these methods in enhancing demographic parity while maintaining high overall performance.
Low GrooveSquid.com (original content) Low Difficulty Summary
Imagine trying to predict what will happen based on some data. Large language models are really good at doing this, but sometimes they don’t do it fairly for everyone. This paper looks at four ways to make sure the predictions are fair and good. They tested these methods using different types of data and large language models. The results show that these methods work well and can be used to make sure predictions are fair.

Keywords

» Artificial intelligence  » Few shot  » Optimization  » Prompt