Fooling Partial Dependence via Data Poisoning

05/26/2021
by   Hubert Baniecki, et al.
0

Many methods have been developed to understand complex predictive models and high expectations are placed on post-hoc model explainability. It turns out that such explanations are not robust nor trustworthy, and they can be fooled. This paper presents techniques for attacking Partial Dependence (plots, profiles, PDP), which are among the most popular methods of explaining any predictive model trained on tabular data. We showcase that PD can be manipulated in an adversarial manner, which is alarming, especially in financial or medical applications where auditability became a must-have trait supporting black-box models. The fooling is performed via poisoning the data to bend and shift explanations in the desired direction using genetic and gradient algorithms. To the best of our knowledge, this is the first work performing attacks on variable dependence explanations. The novel approach of using a genetic algorithm for doing so is highly transferable as it generalizes both ways: in a model-agnostic and an explanation-agnostic manner.

READ FULL TEXT
research
05/04/2020

LIMEtree: Interactively Customisable Explanations Based on Local Surrogate Multi-output Regression Trees

Systems based on artificial intelligence and machine learning models sho...
research
05/28/2021

Do not explain without context: addressing the blind spot of model explanations

The increasing number of regulations and expectations of predictive mach...
research
06/11/2022

Enhancing Explainability of Hyperparameter Optimization via Bayesian Algorithm Execution

Despite all the benefits of automated hyperparameter optimization (HPO),...
research
08/29/2023

Glocal Explanations of Expected Goal Models in Soccer

The expected goal models have gained popularity, but their interpretabil...
research
06/29/2021

Semantic Reasoning from Model-Agnostic Explanations

With the wide adoption of black-box models, instance-based post hoc expl...
research
07/20/2023

Conditional expectation network for SHAP

A very popular model-agnostic technique for explaining predictive models...
research
05/30/2022

Fooling SHAP with Stealthily Biased Sampling

SHAP explanations aim at identifying which features contribute the most ...

Please sign up or login with your details

Forgot password? Click here to reset