A Generative Approach for Mitigating Structural Biases in Natural Language Inference

08/31/2021
by   Dimion Asael, et al.
0

Many natural language inference (NLI) datasets contain biases that allow models to perform well by only using a biased subset of the input, without considering the remainder features. For instance, models are able to make a classification decision by only using the hypothesis, without learning the true relationship between it and the premise. These structural biases lead discriminative models to learn unintended superficial features and to generalize poorly out of the training distribution. In this work, we reformulate the NLI task as a generative task, where a model is conditioned on the biased subset of the input and the label and generates the remaining subset of the input. We show that by imposing a uniform prior, we obtain a provably unbiased model. Through synthetic experiments, we find that this approach is highly robust to large amounts of bias. We then demonstrate empirically on two types of natural bias that this approach leads to fully unbiased models in practice. However, we find that generative models are difficult to train and they generally perform worse than discriminative baselines. We highlight the difficulty of the generative modeling task in the context of NLI as a cause for this worse performance. Finally, by fine-tuning the generative model with a discriminative objective, we reduce the performance gap between the generative model and the discriminative baseline, while allowing for a small amount of bias.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/09/2019

Don't Take the Premise for Granted: Mitigating Artifacts in Natural Language Inference

Natural Language Inference (NLI) datasets often contain hypothesis-only ...
research
10/08/2020

Discriminatively-Tuned Generative Classifiers for Robust Natural Language Inference

While discriminative neural network classifiers are generally preferred,...
research
08/28/2019

Unlearn Dataset Bias in Natural Language Inference by Fitting the Residual

Statistical natural language inference (NLI) models are susceptible to l...
research
10/28/2022

Towards solving model bias in cosmic shear forward modeling

As the volume and quality of modern galaxy surveys increase, so does the...
research
11/03/2020

The Gap on GAP: Tackling the Problem of Differing Data Distributions in Bias-Measuring Datasets

Diagnostic datasets that can detect biased models are an important prere...
research
02/06/2023

Guide the Learner: Controlling Product of Experts Debiasing Method Based on Token Attribution Similarities

Several proposals have been put forward in recent years for improving ou...
research
08/23/2021

Exploring Biases and Prejudice of Facial Synthesis via Semantic Latent Space

Deep learning (DL) models are widely used to provide a more convenient a...

Please sign up or login with your details

Forgot password? Click here to reset