ESD: Expected Squared Difference as a Tuning-Free Trainable Calibration Measure

03/04/2023
by   Hee Suk Yoon, et al.
0

Studies have shown that modern neural networks tend to be poorly calibrated due to over-confident predictions. Traditionally, post-processing methods have been used to calibrate the model after training. In recent years, various trainable calibration measures have been proposed to incorporate them directly into the training process. However, these methods all incorporate internal hyperparameters, and the performance of these calibration objectives relies on tuning these hyperparameters, incurring more computational costs as the size of neural networks and datasets become larger. As such, we present Expected Squared Difference (ESD), a tuning-free (i.e., hyperparameter-free) trainable calibration objective loss, where we view the calibration error from the perspective of the squared difference between the two expectations. With extensive experiments on several architectures (CNNs, Transformers) and datasets, we demonstrate that (1) incorporating ESD into the training improves model calibration in various batch size settings without the need for internal hyperparameter tuning, (2) ESD yields the best-calibrated results compared with previous approaches, and (3) ESD drastically improves the computational costs required for calibration during training due to the absence of internal hyperparameter. The code is publicly accessible at https://github.com/hee-suk-yoon/ESD.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/19/2023

Calibration of Neural Networks

Neural networks solving real-world problems are often required not only ...
research
06/15/2021

Revisiting the Calibration of Modern Neural Networks

Accurate estimation of predictive uncertainty (model calibration) is ess...
research
03/25/2022

A Stitch in Time Saves Nine: A Train-Time Regularizing Loss for Improved Neural Network Calibration

Deep Neural Networks ( DNN s) are known to make overconfident mistakes, ...
research
10/13/2022

A Consistent and Differentiable Lp Canonical Calibration Error Estimator

Calibrated probabilistic classifiers are models whose predicted probabil...
research
09/06/2023

Do We Still Need Non-Maximum Suppression? Accurate Confidence Estimates and Implicit Duplication Modeling with IoU-Aware Calibration

Object detectors are at the heart of many semi- and fully autonomous dec...
research
10/31/2022

A Close Look into the Calibration of Pre-trained Language Models

Pre-trained language models (PLMs) achieve remarkable performance on man...
research
04/28/2023

Online Platt Scaling with Calibeating

We present an online post-hoc calibration method, called Online Platt Sc...

Please sign up or login with your details

Forgot password? Click here to reset