Loading Now

Summary of The Impact Of Geometric Complexity on Neural Collapse in Transfer Learning, by Michael Munn et al.


The Impact of Geometric Complexity on Neural Collapse in Transfer Learning

by Michael Munn, Benoit Dherin, Javier Gonzalvo

First submitted to arxiv on: 24 May 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Many large-scale computer vision and language models owe their success to transfer learning via pre-training foundation models. However, the underlying theoretical framework remains incomplete. To bridge this gap, researchers have introduced metrics like flatness of the loss surface and neural collapse to understand implicit biases during pre-training. This paper delves into the geometric complexity of learned representations as a key mechanism linking these two concepts. Experiments and theory demonstrate how mechanisms affecting geometric complexity influence neural collapse, with implications for better performance on downstream tasks, particularly in few-shot settings.
Low GrooveSquid.com (original content) Low Difficulty Summary
Large-scale computer vision and language models are very good at doing certain tasks. They can even learn from one task and use that knowledge to do other tasks well too. But scientists didn’t have a clear explanation for why this was happening. They came up with some new ways to measure how well these models were pre-trained, like flatness of the loss surface and neural collapse. This paper explores the idea that the way these models learn representations is important for understanding how they transfer knowledge to other tasks. The research shows that making changes to how models learn representations can actually help them do better on new tasks, especially when only given a few examples.

Keywords

» Artificial intelligence  » Few shot  » Transfer learning