Attributions Beyond Neural Networks: The Linear Program Case

06/14/2022
by   Florian Peter Busch, et al.
13

Linear Programs (LPs) have been one of the building blocks in machine learning and have championed recent strides in differentiable optimizers for learning systems. While there exist solvers for even high-dimensional LPs, understanding said high-dimensional solutions poses an orthogonal and unresolved problem. We introduce an approach where we consider neural encodings for LPs that justify the application of attribution methods from explainable artificial intelligence (XAI) designed for neural learning systems. The several encoding functions we propose take into account aspects such as feasibility of the decision space, the cost attached to each input, or the distance to special points of interest. We investigate the mathematical consequences of several XAI methods on said neural LP encodings. We empirically show that the attribution methods Saliency and LIME reveal indistinguishable results up to perturbation levels, and we propose the property of Directedness as the main discriminative criterion between Saliency and LIME on one hand, and a perturbation-based Feature Permutation approach on the other hand. Directedness indicates whether an attribution method gives feature attributions with respect to an increase of that feature. We further notice the baseline selection problem beyond the classical computer vision setting for Integrated Gradients.

READ FULL TEXT

page 6

page 8

research
06/06/2019

Segment Integrated Gradients: Better attributions through regions

Saliency methods can aid understanding of deep neural networks. Recent y...
research
07/12/2018

Maximizing Invariant Data Perturbation with Stochastic Optimization

Feature attribution methods, or saliency maps, are one of the most popul...
research
06/14/2022

Machines Explaining Linear Programs

There has been a recent push in making machine learning models more inte...
research
10/13/2021

When saliency goes off on a tangent: Interpreting Deep Neural Networks with nonlinear saliency maps

A fundamental bottleneck in utilising complex machine learning systems f...
research
07/07/2023

On Formal Feature Attribution and Its Approximation

Recent years have witnessed the widespread use of artificial intelligenc...
research
12/22/2022

Impossibility Theorems for Feature Attribution

Despite a sea of interpretability methods that can produce plausible exp...
research
11/29/2022

Towards More Robust Interpretation via Local Gradient Alignment

Neural network interpretation methods, particularly feature attribution ...

Please sign up or login with your details

Forgot password? Click here to reset