A Framework to Learn with Interpretation

10/19/2020
by   Jayneel Parekh, et al.
0

With increasingly widespread use of deep neural networks in critical decision-making applications, interpretability of these models is becoming imperative. We consider the problem of jointly learning a predictive model and its associated interpretation model. The task of the interpreter is to provide both local and global interpretability about the predictive model in terms of human-understandable high level attribute functions, without any loss of accuracy. This is achieved by a dedicated architecture and well chosen regularization penalties. We seek for a small-size dictionary of attribute functions that take as inputs the outputs of selected hidden layers and whose outputs feed a linear classifier. We impose a high level of conciseness by constraining the activation of a very few attributes for a given input with a real-entropy-based criterion while enforcing fidelity to both inputs and outputs of the predictive model. A major advantage of simultaneous learning is that the predictive neural network benefits from the interpretability constraint as well. We also develop a more detailed pipeline based on some common and novel simple tools to develop understanding about the learnt features. We show on two datasets, MNIST and QuickDraw, their relevance for both global and local interpretability.

READ FULL TEXT

page 7

page 12

page 13

page 14

page 15

page 16

page 17

research
07/09/2018

Supervised Local Modeling for Interpretability

Model interpretability is an increasingly important component of practic...
research
11/08/2020

Unwrapping The Black Box of Deep ReLU Networks: Interpretability, Diagnostics, and Simplification

The deep neural networks (DNNs) have achieved great success in learning ...
research
09/28/2022

Process-guidance improves predictive performance of neural networks for carbon turnover in ecosystems

Despite deep-learning being state-of-the-art for data-driven model predi...
research
11/23/2022

Interpretability of an Interaction Network for identifying H → bb̅ jets

Multivariate techniques and machine learning models have found numerous ...
research
10/31/2022

Hybrid CNN -Interpreter: Interpret local and global contexts for CNN-based Models

Convolutional neural network (CNN) models have seen advanced improvement...
research
05/31/2018

Interpretable Set Functions

We propose learning flexible but interpretable functions that aggregate ...
research
06/12/2022

A Functional Information Perspective on Model Interpretation

Contemporary predictive models are hard to interpret as their deep nets ...

Please sign up or login with your details

Forgot password? Click here to reset