Loading Now

Summary of Characterizing Overfitting in Kernel Ridgeless Regression Through the Eigenspectrum, by Tin Sum Cheng and Aurelien Lucchi and Anastasis Kratsios and David Belius


Characterizing Overfitting in Kernel Ridgeless Regression Through the Eigenspectrum

by Tin Sum Cheng, Aurelien Lucchi, Anastasis Kratsios, David Belius

First submitted to arxiv on: 2 Feb 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Machine Learning (stat.ML)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
In this paper, researchers develop new bounds for kernel matrices, which they use to improve existing test error bounds for kernel ridgeless regression (KRR) in the over-parameterized regime. The new bounds are non-asymptotic and applicable to fixed input dimensions. For kernels with polynomial spectral decay, the results match previous work; for exponential decay, the bounds are novel and non-trivial. The study also explores phenomena such as tempered overfitting and catastrophic overfitting under sub-Gaussian design assumptions, filling a gap in the literature.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper studies how to make predictions more accurate when using kernel ridgeless regression (KRR) with lots of features. It finds that some types of kernel matrices can cause problems if there are too many features. The researchers come up with new rules to help predict how well KRR will work, which can be useful for making machines learn from data.

Keywords

* Artificial intelligence  * Overfitting  * Regression