Loading Now

Summary of Approximations to the Fisher Information Metric Of Deep Generative Models For Out-of-distribution Detection, by Sam Dauncey et al.


Approximations to the Fisher Information Metric of Deep Generative Models for Out-Of-Distribution Detection

by Sam Dauncey, Chris Holmes, Christopher Williams, Fabian Falck

First submitted to arxiv on: 3 Mar 2024

Categories

  • Main: Machine Learning (stat.ML)
  • Secondary: Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper addresses the open problem of out-of-distribution (OOD) detection using likelihood-based deep generative models, specifically score-based diffusion models and variational autoencoders. The authors reproduce seminal work showing that these models consistently infer higher log-likelihoods for OOD data than training data. They propose a new method based on the gradient of a data point with respect to the model parameters, motivated by the intuition that OOD data should have larger gradient norms. This is formalized using the Fisher information metric, which has large absolute diagonal values, leading to chi-square distributed layer-wise gradient norms as features. The authors combine these features for OOD detection and demonstrate their method outperforms the Typicality test for most deep generative models and image dataset pairings.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper is about how we can use special kinds of computer programs called machine learning models to figure out when new data doesn’t fit with what we’ve seen before. These models are really good at guessing what data might look like, but they have a problem: they tend to think new, weird data is actually part of the group they’re used to seeing. The researchers in this paper want to fix that by looking at how the model’s “brain” changes when it sees new data. They came up with a simple way to do this using something called the Fisher information metric. It turns out this works really well, and their method is better than other ways people have tried to solve this problem.

Keywords

* Artificial intelligence  * Likelihood  * Machine learning