Loading Now

Summary of One-step Early Stopping Strategy Using Neural Tangent Kernel Theory and Rademacher Complexity, by Daniel Martin Xavier and Ludovic Chamoin and Jawher Jerray and Laurent Fribourg


One-Step Early Stopping Strategy using Neural Tangent Kernel Theory and Rademacher Complexity

by Daniel Martin Xavier, Ludovic Chamoin, Jawher Jerray, Laurent Fribourg

First submitted to arxiv on: 27 Nov 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Systems and Control (eess.SY)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper proposes an analytical estimation of the optimal early stopping time for neural networks, which involves the initial training error vector and the eigenvalues of the neural tangent kernel. This approach yields an upper bound on the population loss, suitable for underparameterized scenarios where the number of parameters is moderate compared to the number of data. The method is demonstrated using an example of a neural network simulating MPC control of a Van der Pol oscillator.
Low GrooveSquid.com (original content) Low Difficulty Summary
The paper looks at how to stop training a neural network early without overfitting. They develop a way to estimate when it’s best to stop training, based on some initial error and the “neural tangent kernel”. This helps with underparameterized models where there aren’t enough parameters to fit all the data. The method is shown to work well in controlling an oscillator.

Keywords

» Artificial intelligence  » Early stopping  » Neural network  » Overfitting