Loading Now

Summary of Robust Neural Processes For Noisy Data, by Chen Shapira et al.


Robust Neural Processes for Noisy Data

by Chen Shapira, Dan Rosenbaum

First submitted to arxiv on: 3 Nov 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Models that adapt their predictions based on given contexts, known as in-context learning, have become widespread in recent years. We investigate the behavior of such models when data is contaminated by noise using the Neural Processes (NP) framework, which learns a distribution over functions and makes predictions based on context points. Our findings reveal that models performing well on clean data differ from those performing well on noisy data. Specifically, attention-processing models are severely affected by noise, leading to in-context overfitting. To address this issue, we propose a simple method for training NP models that boosts their robustness to noisy data. Experiments on 1D functions and 2D image datasets demonstrate that our approach outperforms all other NP models across various noise levels.
Low GrooveSquid.com (original content) Low Difficulty Summary
In-context learning is when AI models adapt their predictions based on given contexts. But what happens when the data has noise? We looked at how Neural Processes (NP) models work in this situation. NPs learn a distribution over functions and make predictions based on context points. Our results show that models that do well with clean data don’t necessarily do well with noisy data. In fact, attention-processing models are really bad at handling noise. To fix this, we came up with a simple way to train NP models so they’re more robust to noisy data. We tested our method on 1D functions and 2D image datasets and found that it works better than other NPs for all levels of noise.

Keywords

» Artificial intelligence  » Attention  » Overfitting