Learning Interpretable Features via Adversarially Robust Optimization

05/09/2019
by   Ashkan Khakzar, et al.
0

Neural networks are proven to be remarkably successful for classification and diagnosis in medical applications. However, the ambiguity in the decision-making process and the interpretability of the learned features is a matter of concern. In this work, we propose a method for improving the feature interpretability of neural network classifiers. Initially, we propose a baseline convolutional neural network with state of the art performance in terms of accuracy and weakly supervised localization. Subsequently, the loss is modified to integrate robustness to adversarial examples into the training process. In this work, feature interpretability is quantified via evaluating the weakly supervised localization using the ground truth bounding boxes. Interpretability is also visually assessed using class activation maps and saliency maps. The method is applied to NIH ChestX-ray14, the largest publicly available chest x-rays dataset. We demonstrate that the adversarially robust optimization paradigm improves feature interpretability both quantitatively and visually.

READ FULL TEXT

page 3

page 7

research
09/17/2023

FDCNet: Feature Drift Compensation Network for Class-Incremental Weakly Supervised Object Localization

This work addresses the task of class-incremental weakly supervised obje...
research
08/22/2018

Deep multiscale convolutional feature learning for weakly supervised localization of chest pathologies in X-ray images

Localization of chest pathologies in chest X-ray images is a challenging...
research
12/02/2020

Improving Interpretability in Medical Imaging Diagnosis using Adversarial Training

We investigate the influence of adversarial training on the interpretabi...
research
08/01/2020

Eigen-CAM: Class Activation Map using Principal Components

Deep neural networks are ubiquitous due to the ease of developing models...
research
01/27/2022

LAP: An Attention-Based Module for Faithful Interpretation and Knowledge Injection in Convolutional Neural Networks

Despite the state-of-the-art performance of deep convolutional neural ne...
research
11/29/2019

On the Benefits of Attributional Robustness

Interpretability is an emerging area of research in trustworthy machine ...
research
01/26/2018

Weakly Supervised Object Detection with Pointwise Mutual Information

In this work a novel approach for weakly supervised object detection tha...

Please sign up or login with your details

Forgot password? Click here to reset