Learning to Answer by Learning to Ask: Getting the Best of GPT-2 and BERT Worlds

11/06/2019
by   Tassilo Klein, et al.
43

Automatic question generation aims at the generation of questions from a context, with the corresponding answers being sub-spans of the given passage. Whereas, most of the methods mostly rely on heuristic rules to generate questions, more recently also neural network approaches have been proposed. In this work, we propose a variant of the self-attention Transformer network architectures model to generate meaningful and diverse questions. To this end, we propose an easy to use model consisting of the conjunction of the Transformer decoder GPT-2 model with Transformer encoder BERT for the downstream task for question answering. The model is trained in an end-to-end fashion, where the language model is trained to produce a question-answer-aware input representation that facilitates to generate an answer focused question. Our result of neural question generation from text on the SQuAD 1.1 dataset suggests that our method can produce semantically correct and diverse questions. Additionally, we assessed the performance of our proposed method for the downstream task of question answering. The analysis shows that our proposed generation answering collaboration framework relatively improves both tasks and is particularly powerful in the semi-supervised setup. The results further suggest a robust and comparably lean pipeline facilitating question generation in the small-data regime.

READ FULL TEXT

page 2

page 4

research
04/06/2017

Neural Question Generation from Text: A Preliminary Study

Automatic question generation aims to generate questions from a text pas...
research
05/03/2020

Transformer-based End-to-End Question Generation

Question Generation (QG) is an important task in Natural Language Proces...
research
05/10/2022

Learning to Answer Visual Questions from Web Videos

Recent methods for visual question answering rely on large-scale annotat...
research
10/19/2020

Better Distractions: Transformer-based Distractor Generation and Multiple Choice Question Filtering

For the field of education, being able to generate semantically correct ...
research
12/17/2015

Semi-supervised Question Retrieval with Gated Convolutions

Question answering forums are rapidly growing in size with no effective ...
research
09/03/2021

Contextualized Embeddings based Convolutional Neural Networks for Duplicate Question Identification

Question Paraphrase Identification (QPI) is a critical task for large-sc...
research
06/01/2020

When Bert Forgets How To POS: Amnesic Probing of Linguistic Properties and MLM Predictions

A growing body of work makes use of probing in order to investigate the ...

Please sign up or login with your details

Forgot password? Click here to reset