The Real-World-Weight Cross-Entropy Loss Function: Modeling the Costs of Mislabeling

01/03/2020
by   Yaoshiang Ho, et al.
8

In this paper, we propose a new metric to measure goodness-of-fit for classifiers, the Real World Cost function. This metric factors in information about a real world problem, such as financial impact, that other measures like accuracy or F1 do not. This metric is also more directly interpretable for users. To optimize for this metric, we introduce the Real-World- Weight Crossentropy loss function, in both binary and single-label classification variants. Both variants allow direct input of real world costs as weights. For single-label, multicategory classification, our loss function also allows direct penalization of probabilistic false positives, weighted by label, during the training of a machine learning model. We compare the design of our loss function to the binary crossentropy and categorical crossentropy functions, as well as their weighted variants, to discuss the potential for improvement in handling a variety of known shortcomings of machine learning, ranging from imbalanced classes to medical diagnostic error to reinforcement of social bias. We create scenarios that emulate those issues using the MNIST data set and demonstrate empirical results of our new loss function. Finally, we sketch a proof of this function based on Maximum Likelihood Estimation and discuss future directions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/21/2018

Wrapped Loss Function for Regularizing Nonconforming Residual Distributions

Multi-output is essential in machine learning that it might suffer from ...
research
05/27/2018

Metric-Optimized Example Weights

Real-world machine learning applications often have complex test metrics...
research
10/29/2022

Reformulating van Rijsbergen's F_β metric for weighted binary cross-entropy

The separation of performance metrics from gradient based loss functions...
research
06/05/2020

Hierarchical Class-Based Curriculum Loss

Classification algorithms in machine learning often assume a flat label ...
research
01/04/2022

AutoBalance: Optimized Loss Functions for Imbalanced Data

Imbalanced datasets are commonplace in modern machine learning problems....
research
07/25/2022

A novel Deep Learning approach for one-step Conformal Prediction approximation

Deep Learning predictions with measurable confidence are increasingly de...
research
12/13/2020

Demysifying Deep Neural Networks Through Interpretation: A Survey

Modern deep learning algorithms tend to optimize an objective metric, su...

Please sign up or login with your details

Forgot password? Click here to reset