Loading Now

Summary of Temperature Optimization For Bayesian Deep Learning, by Kenyon Ng et al.


Temperature Optimization for Bayesian Deep Learning

by Kenyon Ng, Chris van der Heide, Liam Hodgkinson, Susan Wei

First submitted to arxiv on: 8 Oct 2024

Categories

  • Main: Machine Learning (stat.ML)
  • Secondary: Machine Learning (cs.LG); Computation (stat.CO); Methodology (stat.ME)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper investigates the Cold Posterior Effect (CPE) in Bayesian Deep Learning (BDL), where tempering the posterior distribution improves predictive performance. However, there is no systematic method for selecting the optimal temperature beyond grid search. The authors propose a data-driven approach to estimate the temperature that maximizes test log-predictive density, treating it as a model parameter. They demonstrate that their method performs comparably to grid search at a fraction of the cost across regression and classification tasks. Notably, the paper highlights differing perspectives on CPE between BDL and Generalized Bayes communities, with the former prioritizing predictive performance and the latter emphasizing calibrated uncertainty and robustness.
Low GrooveSquid.com (original content) Low Difficulty Summary
This research looks at how to make computer models better by adjusting a temperature setting. Right now, people are trying different temperatures to see which one works best, but it’s not very efficient. The authors came up with a new way to figure out the best temperature using data from the model itself. They tested their method and found that it worked just as well as the old way, but was much faster and used less computer power. The study also shows how different groups of researchers have different ideas about what makes a good model.

Keywords

» Artificial intelligence  » Classification  » Deep learning  » Grid search  » Regression  » Temperature