Loading Now

Summary of Large Language Models Are Geographically Biased, by Rohin Manvi et al.


Large Language Models are Geographically Biased

by Rohin Manvi, Samar Khanna, Marshall Burke, David Lobell, Stefano Ermon

First submitted to arxiv on: 5 Feb 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI); Computers and Society (cs.CY); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Large Language Models (LLMs) inherently carry biases from their training corpora, which can perpetuate societal harm. To evaluate and achieve fairness and accuracy, understanding LLMs’ knowledge of the world is crucial. Our approach focuses on geography, leveraging ground truth for aspects like culture, race, language, politics, and religion. We demonstrate accurate zero-shot geospatial predictions with strong correlation to ground truth (up to 0.89). However, we also show that LLMs exhibit common biases across various topics, including socioeconomic conditions (e.g., Africa) on sensitive subjective topics (attractiveness, morality, intelligence; up to 0.70 Spearman’s ρ). We introduce a bias score to quantify this and find significant variation across existing LLMs.
Low GrooveSquid.com (original content) Low Difficulty Summary
Large Language Models carry biases from their training data, which can cause problems. To fix these issues, we need to understand what the models know about the world. One way to do this is by looking at how well they can predict geographic information. We found that some models are really good at making predictions (up to 0.89 correlation with the truth). However, we also discovered that many models have biases against certain places or groups of people based on sensitive topics like beauty, morality, and intelligence.

Keywords

* Artificial intelligence  * Zero shot