Loading Now

Summary of Mopd: Mixture-of-prompts Distillation For Vision-language Models, by Yang Chen et al.


MoPD: Mixture-of-Prompts Distillation for Vision-Language Models

by Yang Chen, Shuai Fu, Yu Zhang

First submitted to arxiv on: 26 Dec 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Computation and Language (cs.CL); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper proposes a novel approach called Mixture-of-Prompts Distillation (MoPD) to improve the generalization ability of soft prompt learning methods for vision-language models (VLMs). Existing methods tend to overfit seen classes and perform poorly on unseen classes, which is attributed to the bias in training data. MoPD addresses this issue by transferring knowledge from hand-crafted teacher prompts to learnable student prompts, enhancing performance on unseen classes. The method utilizes a gating network that selects hard prompts for prompt distillation. Experimental results show that MoPD outperforms state-of-the-art baselines, particularly on unseen classes.
Low GrooveSquid.com (original content) Low Difficulty Summary
The paper wants to help machines learn better by creating new ways to teach them. Right now, machines are good at doing what they were trained to do, but not so good at doing something new. The problem is that the training data has a bias towards what the machine was originally taught, making it hard for the machine to learn something new. To solve this issue, the researchers propose a new method called Mixture-of-Prompts Distillation (MoPD). MoPD takes the knowledge from the original teacher and applies it to the student, allowing the machine to learn better even when faced with something new.

Keywords

» Artificial intelligence  » Distillation  » Generalization  » Prompt