Loading Now

Summary of Towards Understanding the Effect Of Pretraining Label Granularity, by Guan Zhe Hong et al.


Towards Understanding the Effect of Pretraining Label Granularity

by Guan Zhe Hong, Yin Cui, Ariel Fuxman, Stanley H. Chan, Enming Luo

First submitted to arxiv on: 29 Mar 2023

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper investigates how different granularities of pre-training labels impact the generalization performance of deep neural networks in image classification tasks. Specifically, it examines the “fine-to-coarse” transfer learning setting where the pre-training label space is more fine-grained than the target problem’s. The authors empirically demonstrate that pre-training on the leaf labels of ImageNet21k yields better transfer results on ImageNet1k compared to coarser granularity levels. They also provide theoretical explanations, showing that fine-grained pre-training enables networks to learn both common and rare features, leading to improved accuracy on challenging test samples. The study further highlights the importance of label hierarchies in pre-training datasets and alignment between pre-training and target labels.
Low GrooveSquid.com (original content) Low Difficulty Summary
This research paper looks at how different levels of detail in training labels affect how well artificial intelligence (AI) models can be used to identify images. They found that when the training labels are more detailed, AI models perform better on a test image set. The researchers also explained why this is happening and showed that when AI models learn both common and rare features, they do better at identifying tricky test images. To make transfer learning work well in practice, it’s important to have a meaningful label hierarchy in the training data and for the training and testing labels to match up.

Keywords

* Artificial intelligence  * Alignment  * Generalization  * Image classification  * Transfer learning