Task-Specific Embeddings for Ante-Hoc Explainable Text Classification

11/30/2022
by   Kishaloy Halder, et al.
0

Current state-of-the-art approaches to text classification typically leverage BERT-style Transformer models with a softmax classifier, jointly fine-tuned to predict class labels of a target task. In this paper, we instead propose an alternative training objective in which we learn task-specific embeddings of text: our proposed objective learns embeddings such that all texts that share the same target class label should be close together in the embedding space, while all others should be far apart. This allows us to replace the softmax classifier with a more interpretable k-nearest-neighbor classification approach. In a series of experiments, we show that this yields a number of interesting benefits: (1) The resulting order induced by distances in the embedding space can be used to directly explain classification decisions. (2) This facilitates qualitative inspection of the training data, helping us to better understand the problem space and identify labelling quality issues. (3) The learned distances to some degree generalize to unseen classes, allowing us to incrementally add new classes without retraining the model. We present extensive experiments which show that the benefits of ante-hoc explainability and incremental learning come at no cost in overall classification accuracy, thus pointing to practical applicability of our proposed approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/08/2015

Mapping Unseen Words to Task-Trained Embedding Spaces

We consider the supervised training setting in which we learn task-speci...
research
07/23/2021

Text Classification and Clustering with Annealing Soft Nearest Neighbor Loss

We define disentanglement as how far class-different data points from ea...
research
06/16/2018

Joint Input-Label Embedding for Neural Text Classification

Neural text classification methods typically treat output classes as cat...
research
11/22/2022

Adaptive Prototypical Networks

Prototypical network for Few shot learning tries to learn an embedding f...
research
02/17/2023

Like a Good Nearest Neighbor: Practical Content Moderation with Sentence Transformers

Modern text classification systems have impressive capabilities but are ...
research
11/04/2019

Metric Learning for Dynamic Text Classification

Traditional text classifiers are limited to predicting over a fixed set ...
research
09/06/2022

Analyzing Transformers in Embedding Space

Understanding Transformer-based models has attracted significant attenti...

Please sign up or login with your details

Forgot password? Click here to reset