Analysis of Softmax Approximation for Deep Classifiers under Input-Dependent Label Noise

03/15/2020
by   Mark Collier, et al.
5

Modelling uncertainty arising from input-dependent label noise is an increasingly important problem. A state-of-the-art approach for classification [Kendall and Gal, 2017] places a normal distribution over the softmax logits, where the mean and variance of this distribution are learned functions of the inputs. This approach achieves impressive empirical performance but lacks theoretical justification. We show that this model is a special case of a well known and theoretically understood model studied in econometrics. Under this view the softmax over the logit distribution is a smooth approximation to an argmax, where the approximation is exact in the zero temperature limit. We further illustrate that the softmax temperature controls a bias-variance trade-off and the optimal point on this trade-off is not always found at 1.0. By tuning the softmax temperature, we achieve improved performance on well known image classification benchmarks with controlled label noise. For image segmentation, where input-dependent label noise naturally arises, we show that tuning the temperature increases the mean IoU on the PASCAL VOC and Cityscapes datasets by more than 1 that does not model this noise source.

READ FULL TEXT

page 9

page 16

page 17

page 18

research
06/02/2022

Robustness to Label Noise Depends on the Shape of the Noise Distribution in Feature Space

Machine learning classifiers have been demonstrated, both empirically an...
research
05/19/2021

Correlated Input-Dependent Label Noise in Large-Scale Image Classification

Large scale image classification datasets often contain noisy labels. We...
research
05/19/2017

Softmax Q-Distribution Estimation for Structured Prediction: A Theoretical Interpretation for RAML

Reward augmented maximum likelihood (RAML), a simple and effective learn...
research
12/02/2018

Revisiting the Softmax Bellman Operator: Theoretical Properties and Practical Benefits

The softmax function has been primarily employed in reinforcement learni...
research
04/11/2023

r-softmax: Generalized Softmax with Controllable Sparsity Rate

Nowadays artificial neural network models achieve remarkable results in ...
research
02/01/2023

Learning Prototype Classifiers for Long-Tailed Recognition

The problem of long-tailed recognition (LTR) has received attention in r...
research
12/12/2016

Analysis and Optimization of Loss Functions for Multiclass, Top-k, and Multilabel Classification

Top-k error is currently a popular performance measure on large scale im...

Please sign up or login with your details

Forgot password? Click here to reset