Addressing Limited Data for Textual Entailment Across Domains

06/08/2016
by   Chaitanya Shivade, et al.
0

We seek to address the lack of labeled data (and high cost of annotation) for textual entailment in some domains. To that end, we first create (for experimental purposes) an entailment dataset for the clinical domain, and a highly competitive supervised entailment system, ENT, that is effective (out of the box) on two domains. We then explore self-training and active learning strategies to address the lack of labeled data. With self-training, we successfully exploit unlabeled data to improve over ENT by 15 newswire domain, and 13 active learning experiments demonstrate that we can match (and even beat) ENT using only 6.6 the training data in the newswire domain.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/26/2019

The Use of Unlabeled Data versus Labeled Data for Stopping Active Learning for Text Classification

Annotation of training data is the major bottleneck in the creation of t...
research
01/31/2023

Iterative Loop Learning Combining Self-Training and Active Learning for Domain Adaptive Semantic Segmentation

Recently, self-training and active learning have been proposed to allevi...
research
12/16/2021

ATM: An Uncertainty-aware Active Self-training Framework for Label-efficient Text Classification

Despite the great success of pre-trained language models (LMs) in many n...
research
07/23/2019

Dr.Quad at MEDIQA 2019: Towards Textual Inference and Question Entailment using contextualized representations

This paper presents the submissions by Team Dr.Quad to the ACL-BioNLP 20...
research
10/06/2020

Universal Natural Language Processing with Limited Annotations: Try Few-shot Textual Entailment as a Start

A standard way to address different NLP problems is by first constructin...
research
05/04/2023

Multi-Domain Learning From Insufficient Annotations

Multi-domain learning (MDL) refers to simultaneously constructing a mode...
research
04/15/2021

Adaptive Active Learning for Coreference Resolution

Training coreference resolution models require comprehensively labeled d...

Please sign up or login with your details

Forgot password? Click here to reset