Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources

10/30/2016
by   Jason Naradowsky, et al.
0

In order to extract event information from text, a machine reading model must learn to accurately read and interpret the ways in which that information is expressed. But it must also, as the human reader must, aggregate numerous individual value hypotheses into a single coherent global analysis, applying global constraints which reflect prior knowledge of the domain. In this work we focus on the task of extracting plane crash event information from clusters of related news articles whose labels are derived via distant supervision. Unlike previous machine reading work, we assume that while most target values will occur frequently in most clusters, they may also be missing or incorrect. We introduce a novel neural architecture to explicitly model the noisy nature of the data and to deal with these aforementioned learning issues. Our models are trained end-to-end and achieve an improvement of more than 12.1 F_1 over previous work, despite using far less linguistic annotation. We apply factor graph constraints to promote more coherent event analyses, with belief propagation inference formulated within the transitions of a recurrent neural network. We show this technique additionally improves maximum F_1 by up to 2.8 points, resulting in a relative improvement of 50% over the previous state-of-the-art.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2019

Leveraging Knowledge Bases in LSTMs for Improving Machine Reading

This paper focuses on how to take advantage of external knowledge bases ...
research
09/26/2019

Read, Attend and Comment: A Deep Architecture for Automatic News Comment Generation

Automatic news comment generation is beneficial for real applications bu...
research
09/15/2020

Domain Knowledge Empowered Structured Neural Net for End-to-End Event Temporal Relation Extraction

Extracting event temporal relations is a critical task for information e...
research
09/18/2017

Sequence to Sequence Learning for Event Prediction

This paper presents an approach to the task of predicting an event descr...
research
03/20/2019

Neural Speed Reading with Structural-Jump-LSTM

Recurrent neural networks (RNNs) can model natural language by sequentia...
research
11/05/2020

Improving Event Duration Prediction via Time-aware Pre-training

End-to-end models in NLP rarely encode external world knowledge about le...
research
02/23/2019

Evidence Sentence Extraction for Machine Reading Comprehension

Recently remarkable success has been achieved in machine reading compreh...

Please sign up or login with your details

Forgot password? Click here to reset