Exploring Lexical Irregularities in Hypothesis-Only Models of Natural Language Inference

by   Qingyuan Hu, et al.

Natural Language Inference (NLI) or Recognizing Textual Entailment (RTE) is the task of predicting the entailment relation between a pair of sentences (premise and hypothesis). This task has been described as a valuable testing ground for the development of semantic representations, and is a key component in natural language understanding evaluation benchmarks. Models that understand entailment should encode both, the premise and the hypothesis. However, experiments by Poliak et al. revealed a strong preference of these models towards patterns observed only in the hypothesis, based on a 10 dataset comparison. Their results indicated the existence of statistical irregularities present in the hypothesis that bias the model into performing competitively with the state of the art. While recast datasets provide large scale generation of NLI instances due to minimal human intervention, the papers that generate them do not provide fine-grained analysis of the potential statistical patterns that can bias NLI models. In this work, we analyze hypothesis-only models trained on one of the recast datasets provided in Poliak et al. for word-level patterns. Our results indicate the existence of potential lexical biases that could contribute to inflating the model performance.


page 1

page 2

page 3

page 4


Textual Entailment Recognition with Semantic Features from Empirical Text Representation

Textual entailment recognition is one of the basic natural language unde...

Language Models for Lexical Inference in Context

Lexical inference in context (LIiC) is the task of recognizing textual e...

Performance Impact Caused by Hidden Bias of Training Data for Recognizing Textual Entailment

The quality of training data is one of the crucial problems when a learn...

Hypothesis Only Baselines in Natural Language Inference

We propose a hypothesis only baseline for diagnosing Natural Language In...

NLVR2 Visual Bias Analysis

NLVR2 (Suhr et al., 2019) was designed to be robust for language bias th...

Looking at the Overlooked: An Analysis on the Word-Overlap Bias in Natural Language Inference

It has been shown that NLI models are usually biased with respect to the...

HypoNLI: Exploring the Artificial Patterns of Hypothesis-only Bias in Natural Language Inference

Many recent studies have shown that for models trained on datasets for n...

Please sign up or login with your details

Forgot password? Click here to reset