Linearity of Relation Decoding in Transformer Language Models

08/17/2023
by   Evan Hernandez, et al.
0

Much of the knowledge encoded in transformer language models (LMs) may be expressed in terms of relations: relations between words and their synonyms, entities and their attributes, etc. We show that, for a subset of relations, this computation is well-approximated by a single linear transformation on the subject representation. Linear relation representations may be obtained by constructing a first-order approximation to the LM from a single prompt, and they exist for a variety of factual, commonsense, and linguistic relations. However, we also identify many cases in which LM predictions capture relational knowledge accurately, but this knowledge is not linearly encoded in their representations. Our results thus reveal a simple, interpretable, but heterogeneously deployed knowledge representation strategy in transformer LMs.

READ FULL TEXT

page 6

page 7

page 9

page 17

page 19

research
06/27/2022

Analyzing Encoded Concepts in Transformer Language Models

We propose a novel framework ConceptX, to analyze how latent concepts ar...
research
07/06/2023

Extracting Multi-valued Relations from Language Models

The widespread usage of latent language representations via pre-trained ...
research
04/28/2023

Dissecting Recall of Factual Associations in Auto-Regressive Language Models

Transformer-based language models (LMs) are known to capture factual kno...
research
09/04/2017

Compositional Approaches for Representing Relations Between Words: A Comparative Study

Identifying the relations that exist between words (or entities) is impo...
research
03/01/2023

Competence-Based Analysis of Language Models

Despite the recent success of large pretrained language models (LMs) on ...
research
01/06/2023

Topics as Entity Clusters: Entity-based Topics from Language Models and Graph Neural Networks

Topic models aim to reveal the latent structure behind a corpus, typical...
research
10/08/2021

Text analysis and deep learning: A network approach

Much information available to applied researchers is contained within wr...

Please sign up or login with your details

Forgot password? Click here to reset