Loading Now

Summary of Reconstructing Training Data From Real World Models Trained with Transfer Learning, by Yakir Oz et al.


Reconstructing Training Data From Real World Models Trained with Transfer Learning

by Yakir Oz, Gilad Yehudai, Gal Vardi, Itai Antebi, Michal Irani, Niv Haim

First submitted to arxiv on: 22 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Artificial Intelligence (cs.AI); Cryptography and Security (cs.CR); Computer Vision and Pattern Recognition (cs.CV)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
A novel approach is proposed to reconstruct training data from trained classifiers, enabling data reconstruction in realistic settings for models trained on high-resolution images. The method adapts the reconstruction scheme of arXiv:2206.07758 to real-world scenarios, targeting models trained via transfer learning over image embeddings of large pre-trained models like DINO-ViT and CLIP. Data reconstruction is employed in the embedding space, showcasing its applicability beyond visual data. A novel clustering-based method is introduced to identify good reconstructions from thousands of candidates, improving on previous works that relied on knowledge of the training set to identify good reconstructed images.
Low GrooveSquid.com (original content) Low Difficulty Summary
Imagine a way to see what your trained AI model learned from, even if it was trained on super high-quality pictures! This new approach makes it possible to do just that. It uses clever tricks from another study (arXiv:2206.07758) and applies them to real-world situations where models are trained using big pre-trained models like DINO-ViT and CLIP. The cool thing is that this method works not just for pictures, but for any kind of data that can be embedded into a special space. It also introduces a new way to pick the best reconstructed versions from many possible choices.

Keywords

» Artificial intelligence  » Clustering  » Embedding space  » Transfer learning  » Vit