On Calibration of Modern Neural Networks
Chuan Guo, Geoff Pleiss, Yu Sun, Kilian Q. Weinberger
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/gpleiss/temperature_scalingOfficialIn paperpytorch★ 0
- github.com/aigen/df-posthoc-calibrationnone★ 41
- github.com/anonwhymoos/connectivityjax★ 18
- github.com/sleep3r/garrusnone★ 3
- github.com/johntd54/stanford_carpytorch★ 0
- github.com/AnanyaKumar/verified_calibrationtf★ 0
- github.com/sirius8050/Expected-Calibration-Errornone★ 0
- github.com/saurabhgarg1996/calibrationpytorch★ 0
- github.com/Andreas12321/Est-Cert-Finaltf★ 0
- github.com/cpark321/bayesian-neural-networkspytorch★ 0
Abstract
Confidence calibration -- the problem of predicting probability estimates representative of the true correctness likelihood -- is important for classification models in many applications. We discover that modern neural networks, unlike those from a decade ago, are poorly calibrated. Through extensive experiments, we observe that depth, width, weight decay, and Batch Normalization are important factors influencing calibration. We evaluate the performance of various post-processing calibration methods on state-of-the-art architectures with image and document classification datasets. Our analysis and experiments not only offer insights into neural network learning, but also provide a simple and straightforward recipe for practical settings: on most datasets, temperature scaling -- a single-parameter variant of Platt Scaling -- is surprisingly effective at calibrating predictions.