Loading Now

Summary of Downstream Transfer Attack: Adversarial Attacks on Downstream Models with Pre-trained Vision Transformers, by Weijie Zheng et al.


Downstream Transfer Attack: Adversarial Attacks on Downstream Models with Pre-trained Vision Transformers

by Weijie Zheng, Xingjun Ma, Hanxun Huang, Zuxuan Wu, Yu-Gang Jiang

First submitted to arxiv on: 3 Aug 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper investigates the vulnerability of pre-trained vision transformers (ViTs) to sample-wise transfer attacks. The researchers propose a novel attack method called Downstream Transfer Attack (DTA), which leverages a pre-trained ViT model to craft adversarial examples and then applies them to fine-tuned models on downstream datasets. DTA identifies vulnerable layers in the pre-trained model using a cosine similarity loss, allowing it to generate highly transferable attacks. The study demonstrates the effectiveness of DTA across 10 diverse downstream datasets, achieving an average attack success rate (ASR) exceeding 90%, outperforming existing methods by a significant margin. When used with adversarial training, the generated adversarial examples improve the model’s robustness to different downstream transfer attacks.
Low GrooveSquid.com (original content) Low Difficulty Summary
This research looks at how well pre-trained artificial intelligence models can be tricked into making mistakes. These models are called vision transformers, or ViTs for short. The researchers found that these models can be fooled by adding small changes to the input images. They created a new way to do this called Downstream Transfer Attack (DTA). DTA is very good at finding and exploiting weaknesses in the pre-trained models. In tests with 10 different types of data, DTA was able to trick the models into making mistakes more than 90% of the time! When used together with special training, these tricks can actually make the models better at handling similar attacks in the future.

Keywords

» Artificial intelligence  » Cosine similarity  » Vit