Loading Now

Summary of Interlude: Interactions Between Labeled and Unlabeled Data to Enhance Semi-supervised Learning, by Zhe Huang et al.


InterLUDE: Interactions between Labeled and Unlabeled Data to Enhance Semi-Supervised Learning

by Zhe Huang, Xiaowei Yu, Dajiang Zhu, Michael C. Hughes

First submitted to arxiv on: 15 Mar 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Medium Difficulty summary: Semi-supervised learning (SSL) aims to boost task performance by utilizing both labeled and unlabeled data. Traditional SSL image classification methods primarily focus on optimizing a loss that combines supervised classification objectives with regularization terms derived from unlabeled data. However, this formulation disregards the potential interaction between labeled and unlabeled images. To address this limitation, we introduce InterLUDE, a novel approach consisting of two parts: embedding fusion, which interpolates between labeled and unlabeled embeddings to enhance representation learning, and a new loss function grounded in consistency regularization, aiming to minimize discrepancies in model predictions between labeled and unlabeled inputs. Our method demonstrates clear benefits on standard closed-set SSL benchmarks and a medical SSL task with an uncurrated unlabeled set. Specifically, on the STL-10 dataset with only 40 labels, InterLUDE achieves an impressive error rate of 3.2%, outperforming previous methods with an error rate of 14.9%.
Low GrooveSquid.com (original content) Low Difficulty Summary
Low Difficulty summary: This paper is about a new way to improve machine learning using both labeled and unlabeled data. Traditional methods mix the two types of data without considering how they interact. The authors introduce InterLUDE, which has two main parts: one that combines information from labeled and unlabeled images to create better representations, and another that tries to make sure the model makes consistent predictions about labeled and unlabeled images. This new approach works well on several tests, including a medical task where it outperforms previous methods.

Keywords

* Artificial intelligence  * Classification  * Embedding  * Image classification  * Loss function  * Machine learning  * Regularization  * Representation learning  * Semi supervised  * Supervised