Deep Learning for Case-Based Reasoning through Prototypes: A Neural Network that Explains Its Predictions

10/13/2017
by   Oscar Li, et al.
0

Deep neural networks are widely used for classification. These deep models often suffer from a lack of interpretability -- they are particularly difficult to understand because of their non-linear nature. As a result, neural networks are often treated as "black box" models, and in the past, have been trained purely to optimize the accuracy of predictions. In this work, we create a novel network architecture for deep learning that naturally explains its own reasoning for each prediction. This architecture contains an autoencoder and a special prototype layer, where each unit of that layer stores a weight vector that resembles an encoded training input. The encoder of the autoencoder allows us to do comparisons within the latent space, while the decoder allows us to visualize the learned prototypes. The training objective has four terms: an accuracy term, a term that encourages every prototype to be similar to at least one encoded input, a term that encourages every encoded input to be close to at least one prototype, and a term that encourages faithful reconstruction by the autoencoder. The distances computed in the prototype layer are used as part of the classification process. Since the prototypes are learned during training, the learned network naturally comes with explanations for each prediction, and the explanations are loyal to what the network actually computes.

READ FULL TEXT

page 5

page 6

research
12/02/2021

ProtGNN: Towards Self-Explaining Graph Neural Networks

Despite the recent progress in Graph Neural Networks (GNNs), it remains ...
research
06/05/2023

Interpretable Alzheimer's Disease Classification Via a Contrastive Diffusion Autoencoder

In visual object classification, humans often justify their choices by c...
research
04/02/2019

A PCA-like Autoencoder

An autoencoder is a neural network which data projects to and from a low...
research
07/31/2022

INSightR-Net: Interpretable Neural Network for Regression using Similarity-based Comparisons to Prototypical Examples

Convolutional neural networks (CNNs) have shown exceptional performance ...
research
04/11/2022

ProtoTEx: Explaining Model Decisions with Prototype Tensors

We present ProtoTEx, a novel white-box NLP classification architecture b...
research
05/31/2022

Concept-level Debugging of Part-Prototype Networks

Part-prototype Networks (ProtoPNets) are concept-based classifiers desig...
research
06/16/2023

Prototype Learning for Explainable Regression

The lack of explainability limits the adoption of deep learning models i...

Please sign up or login with your details

Forgot password? Click here to reset