Smoothing Entailment Graphs with Language Models

07/30/2022
by   Nick McKenna, et al.
2

The diversity and Zipfian frequency distribution of natural language predicates in corpora leads to sparsity when learning Entailment Graphs. As symbolic models for natural language inference, an EG cannot recover if missing a novel premise or hypothesis at test-time. In this paper we approach the problem of vertex sparsity by introducing a new method of graph smoothing, using a Language Model to find the nearest approximations of missing predicates. We improve recall by 25.1 and 16.3 absolute percentage points on two difficult directional entailment datasets while exceeding average precision, and show a complementarity with other improvements to edge sparsity. We further analyze language model embeddings and discuss why they are naturally suitable for premise-smoothing, but not hypothesis-smoothing. Finally, we formalize a theory for smoothing a symbolic inference method by constructing transitive chains to smooth both the premise and hypothesis.

READ FULL TEXT
research
09/26/2022

Entailment Semantics Can Be Extracted from an Ideal Language Model

Language models are often trained on text alone, without additional grou...
research
04/16/2021

Multivalent Entailment Graphs for Question Answering

Drawing inferences between open-domain natural language predicates is a ...
research
06/07/2023

From the One, Judge of the Whole: Typed Entailment Graph Construction with Predicate Generation

Entailment Graphs (EGs) have been constructed based on extracted corpora...
research
10/10/2022

Language Models Are Poor Learners of Directional Inference

We examine LMs' competence of directional predicate entailments by super...
research
05/02/2018

Hypothesis Only Baselines in Natural Language Inference

We propose a hypothesis only baseline for diagnosing Natural Language In...
research
06/01/2023

AMR4NLI: Interpretable and robust NLI measures from semantic graphs

The task of natural language inference (NLI) asks whether a given premis...
research
01/28/2022

Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval

Retrieval-based language models (R-LM) model the probability of natural ...

Please sign up or login with your details

Forgot password? Click here to reset