Loading Now

Summary of Modeling Caption Diversity in Contrastive Vision-language Pretraining, by Samuel Lavoie et al.


Modeling Caption Diversity in Contrastive Vision-Language Pretraining

by Samuel Lavoie, Polina Kirichenko, Mark Ibrahim, Mahmoud Assran, Andrew Gordon Wilson, Aaron Courville, Nicolas Ballas

First submitted to arxiv on: 30 Apr 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Contrastive Language Pretraining (CLIP) models an image and its caption as a single vector, limiting representation diversity. This paper introduces Llip, Latent Language Image Pretraining, which generates diverse captions for an image by conditioning on text-derived information. Llip’s vision encoder outputs visual features mixed with textual information to produce a final representation. The proposed method outperforms non-contextualized baselines like CLIP and SigLIP on various tasks, even with large-scale encoders. Notably, Llip improves zero-shot classification by 2.9% and achieves a top-1 accuracy of 83.5% on ImageNet, surpassing a similarly sized CLIP by 1.4%. The method also demonstrates improved performance on zero-shot retrieval on MS-COCO by 6.0%. A comprehensive analysis of Llip’s components shows that it leads to richer visual representations.
Low GrooveSquid.com (original content) Low Difficulty Summary
Imagine you want to describe an image in many different ways. This paper proposes a new way to do this, called Llip. Unlike other methods that try to match images with captions by using a single vector, Llip tries to capture the many different ways people could describe an image. The authors show that Llip is better than other methods at tasks like classifying images and finding similar images. They also show that Llip can improve how well computers understand what’s in an image.

Keywords

» Artificial intelligence  » Classification  » Encoder  » Pretraining  » Zero shot