Towards A Holistic View of Bias in Machine Learning: Bridging Algorithmic Fairness and Imbalanced Learning

07/13/2022
by   Damien Dablain, et al.
0

Machine learning (ML) is playing an increasingly important role in rendering decisions that affect a broad range of groups in society. ML models inform decisions in criminal justice, the extension of credit in banking, and the hiring practices of corporations. This posits the requirement of model fairness, which holds that automated decisions should be equitable with respect to protected features (e.g., gender, race, or age) that are often under-represented in the data. We postulate that this problem of under-representation has a corollary to the problem of imbalanced data learning. This class imbalance is often reflected in both classes and protected features. For example, one class (those receiving credit) may be over-represented with respect to another class (those not receiving credit) and a particular group (females) may be under-represented with respect to another group (males). A key element in achieving algorithmic fairness with respect to protected groups is the simultaneous reduction of class and protected group imbalance in the underlying training data, which facilitates increases in both model accuracy and fairness. We discuss the importance of bridging imbalanced learning and group fairness by showing how key concepts in these fields overlap and complement each other; and propose a novel oversampling algorithm, Fair Oversampling, that addresses both skewed class distributions and protected features. Our method: (i) can be used as an efficient pre-processing algorithm for standard ML algorithms to jointly address imbalance and group equity; and (ii) can be combined with fairness-aware learning algorithms to improve their robustness to varying levels of class imbalance. Additionally, we take a step toward bridging the gap between fairness and imbalanced learning with a new metric, Fair Utility, that combines balanced accuracy with fairness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/03/2020

FAE: A Fairness-Aware Ensemble Framework

Automated decision making based on big data and machine learning (ML) al...
research
08/01/2023

Fair Models in Credit: Intersectional Discrimination and the Amplification of Inequity

The increasing usage of new data sources and machine learning (ML) techn...
research
05/23/2023

Fair Oversampling Technique using Heterogeneous Clusters

Class imbalance and group (e.g., race, gender, and age) imbalance are ac...
research
04/30/2022

Fair Feature Subset Selection using Multiobjective Genetic Algorithm

The feature subset selection problem aims at selecting the relevant subs...
research
01/25/2022

Beyond the Frontier: Fairness Without Accuracy Loss

Notions of fair machine learning that seek to control various kinds of e...
research
06/30/2021

Unaware Fairness: Hierarchical Random Forest for Protected Classes

Procedural fairness has been a public concern, which leads to controvers...
research
01/20/2023

Fair Credit Scorer through Bayesian Approach

Machine learning currently plays an increasingly important role in peopl...

Please sign up or login with your details

Forgot password? Click here to reset