Learn to explain yourself, when you can: Equipping Concept Bottleneck Models with the ability to abstain on their concept predictions

11/21/2022
by   Joshua Lockhart, et al.
0

The Concept Bottleneck Models (CBMs) of Koh et al. [2020] provide a means to ensure that a neural network based classifier bases its predictions solely on human understandable concepts. The concept labels, or rationales as we refer to them, are learned by the concept labeling component of the CBM. Another component learns to predict the target classification label from these predicted concept labels. Unfortunately, these models are heavily reliant on human provided concept labels for each datapoint. To enable CBMs to behave robustly when these labels are not readily available, we show how to equip them with the ability to abstain from predicting concepts when the concept labeling component is uncertain. In other words, our model learns to provide rationales for its predictions, but only whenever it is sure the rationale is correct.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/07/2022

Towards learning to explain with concept bottleneck models: mitigating information leakage

Concept bottleneck models perform classification by first predicting whi...
research
12/14/2022

Interactive Concept Bottleneck Models

Concept bottleneck models (CBMs) (Koh et al. 2020) are interpretable neu...
research
06/02/2023

Probabilistic Concept Bottleneck Models

Interpretable models are designed to make decisions in a human-interpret...
research
12/11/2021

Multimodal neural networks better explain multivoxel patterns in the hippocampus

The human hippocampus possesses "concept cells", neurons that fire when ...
research
06/21/2022

Automatic Concept Extraction for Concept Bottleneck-based Video Classification

Recent efforts in interpretable deep learning models have shown that con...
research
08/25/2023

Learning to Intervene on Concept Bottlenecks

While traditional deep learning models often lack interpretability, conc...
research
02/28/2023

A Closer Look at the Intervention Procedure of Concept Bottleneck Models

Concept bottleneck models (CBMs) are a class of interpretable neural net...

Please sign up or login with your details

Forgot password? Click here to reset