Loading Now

Summary of Bdetclip: Multimodal Prompting Contrastive Test-time Backdoor Detection, by Yuwei Niu et al.


BDetCLIP: Multimodal Prompting Contrastive Test-Time Backdoor Detection

by Yuwei Niu, Shuo He, Qi Wei, Zongyu Wu, Feng Liu, Lei Feng

First submitted to arxiv on: 24 May 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper proposes a novel computationally efficient method for detecting backdoored CLIP models at the inference stage. Recent research has shown that multimodal contrastive learning can be vulnerable to backdoor attacks, which involve inserting triggers into pre-trained models. Existing defense methods focus on either the pre-training or fine-tuning stages, but these approaches require numerous parameter updates and are computationally expensive. The proposed method, called BDetCLIP, uses a language model (e.g., GPT-4) to produce class-related description texts and perturbed random texts, which are then used to compute cosine similarity with images. This distribution difference is used as the criterion to detect backdoor samples. Experimental results show that BDetCLIP outperforms state-of-the-art methods in both effectiveness and efficiency.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper tries to fix a problem with AI models called “backdoored CLIP”. These models are trained on fake data, which can make them do things they shouldn’t when used later. The researchers found that these backdoor attacks can be very successful. They also found that the way we usually defend against this kind of attack is too slow and uses a lot of computing power. To solve this problem, they came up with a new method called BDetCLIP. This method uses special texts to trick the AI model into revealing when it’s being attacked. The results show that their method works better than others at catching backdoor attacks while also being faster.

Keywords

» Artificial intelligence  » Cosine similarity  » Fine tuning  » Gpt  » Inference  » Language model