Natural Language Descriptions of Deep Visual Features

01/26/2022
by   Evan Hernandez, et al.
2

Some neurons in deep networks specialize in recognizing highly specific perceptual, structural, or semantic features of inputs. In computer vision, techniques exist for identifying neurons that respond to individual concept categories like colors, textures, and object classes. But these techniques are limited in scope, labeling only a small subset of neurons and behaviors in any network. Is a richer characterization of neuron-level computation possible? We introduce a procedure (called MILAN, for mutual-information-guided linguistic annotation of neurons) that automatically labels neurons with open-ended, compositional, natural language descriptions. Given a neuron, MILAN generates a description by searching for a natural language string that maximizes pointwise mutual information with the image regions in which the neuron is active. MILAN produces fine-grained descriptions that capture categorical, relational, and logical structure in learned features. These descriptions obtain high agreement with human-generated feature descriptions across a diverse set of model architectures and tasks, and can aid in understanding and controlling learned models. We highlight three applications of natural language neuron descriptions. First, we use MILAN for analysis, characterizing the distribution and importance of neurons selective for attribute, category, and relational information in vision models. Second, we use MILAN for auditing, surfacing neurons sensitive to protected categories like race and gender in models trained on datasets intended to obscure these features. Finally, we use MILAN for editing, improving robustness in an image classifier by deleting neurons sensitive to text features spuriously correlated with class labels.

READ FULL TEXT

page 4

page 7

page 8

page 9

page 14

page 15

page 19

page 20

research
06/24/2020

Compositional Explanations of Neurons

We describe a procedure for explaining neurons in deep representations b...
research
04/23/2022

CLIP-Dissect: Automatic Description of Neuron Representations in Deep Vision Networks

In this paper, we propose CLIP-Dissect, a new technique to automatically...
research
09/04/2023

DeViL: Decoding Vision features into Language

Post-hoc explanation methods have often been criticised for abstracting ...
research
11/01/2022

Higher-order mutual information reveals synergistic sub-networks for multi-neuron importance

Quantifying which neurons are important with respect to the classificati...
research
11/01/2017

Learning with Latent Language

The named concepts and compositional operators present in natural langua...
research
09/19/2023

Rigorously Assessing Natural Language Explanations of Neurons

Natural language is an appealing medium for explaining how large languag...
research
01/28/2022

Summarizing Differences between Text Distributions with Natural Language

How do two distributions of texts differ? Humans are slow at answering t...

Please sign up or login with your details

Forgot password? Click here to reset