Loading Now

Summary of Epistemic Uncertainty Quantification For Pre-trained Neural Network, by Hanjing Wang and Qiang Ji


Epistemic Uncertainty Quantification For Pre-trained Neural Network

by Hanjing Wang, Qiang Ji

First submitted to arxiv on: 15 Apr 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Computer Vision and Pattern Recognition (cs.CV)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed paper addresses epistemic uncertainty quantification (UQ) for pre-trained non-Bayesian models, which can be applied to any network architecture or training technique without requiring the original training data or model modifications. The authors introduce a gradient-based approach to quantify epistemic UQ by analyzing the gradients of outputs relative to model parameters, indicating necessary model adjustments to accurately represent inputs. This method is theoretically guaranteed and improves upon existing methods through class-specific weights for integrating gradients and emphasizing distinct contributions from neural network layers. The authors also enhance UQ accuracy by combining gradient and perturbation methods to refine gradients. The proposed approach outperforms current state-of-the-art UQ methods in out-of-distribution detection, uncertainty calibration, and active learning tasks.
Low GrooveSquid.com (original content) Low Difficulty Summary
For any pre-trained model, the new method can identify where it lacks knowledge without requiring the original training data or model modifications. This is achieved by analyzing the gradients of outputs relative to model parameters, which indicates necessary adjustments for accurate representation. The approach is theoretically guaranteed and improves upon existing methods through class-specific weights and emphasizing neural network layers. The proposed method is evaluated on out-of-distribution detection, uncertainty calibration, and active learning tasks, showing its superiority over current state-of-the-art UQ methods.

Keywords

» Artificial intelligence  » Active learning  » Neural network