Loading Now

Summary of Fairpivara: Reducing and Assessing Biases in Clip-based Multimodal Models, by Diego A. B. Moreira et al.


FairPIVARA: Reducing and Assessing Biases in CLIP-Based Multimodal Models

by Diego A. B. Moreira, Alef Iury Ferreira, Jhessica Silva, Gabriel Oliveira dos Santos, Luiz Pereira, João Medrado Gondim, Gustavo Bonil, Helena Maia, Nádia da Silva, Simone Tiemi Hashiguti, Jefersson A. dos Santos, Helio Pedrini, Sandra Avila

First submitted to arxiv on: 28 Sep 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Despite significant advancements in vision-language models, there is a lack of research on their ethical implications. These models often rely on extensive training data from imbalanced datasets and raise concerns about biases. The CLIP model, initially trained in English, can be fine-tuned for other languages, introducing new biases. A CLIP-based Portuguese model, CAPIVARA, has shown strong zero-shot performance. This paper evaluates four types of discriminatory practices within visual-language models and proposes FairPIVARA, a method to reduce biases by removing feature embedding dimensions. Applying FairPIVARA reduces observed biases up to 98%, promoting a balanced word distribution. The authors’ model and code are available on GitHub.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper looks at the ethical concerns surrounding vision-language models. These models need a lot of data to work well, but this data can be unbalanced and cause problems. When we fine-tune these models for other languages, it can introduce new biases. The researchers tested four different types of bias in these models and found that they all have discriminatory practices. To fix this, they developed FairPIVARA, a method to reduce these biases by changing how the model processes information. This helps to make the model more fair and balanced.

Keywords

» Artificial intelligence  » Embedding  » Zero shot