Neural Transfer Learning with Transformers for Social Science Text Analysis

02/03/2021
by   Sandra Wankmüller, et al.
48

During the last years, there have been substantial increases in the prediction performances of natural language processing models on text-based supervised learning tasks. Especially deep learning models that are based on the Transformer architecture (Vaswani et al., 2017) and are used in a transfer learning setting have contributed to this development. As Transformer-based models for transfer learning have the potential to achieve higher prediction accuracies with relatively few training data instances, they are likely to benefit social scientists that seek to have as accurate as possible text-based measures but only have limited resources for annotating training data. To enable social scientists to leverage these potential benefits for their research, this paper explains how these methods work, why they might be advantageous, and what their limitations are. Additionally, three Transformer-based models for transfer learning, BERT (Devlin et al., 2019), RoBERTa (Liu et al., 2019), and the Longformer (Beltagy et al., 2020), are compared to conventional machine learning algorithms on three social science applications. Across all evaluated tasks, textual styles, and training data set sizes, the conventional models are consistently outperformed by transfer learning with Transformer-based models, thereby demonstrating the potential benefits these models can bring to text-based social science research.

READ FULL TEXT

page 26

page 27

page 34

page 35

research
08/17/2022

Transformer Encoder for Social Science

High-quality text data has become an important data source for social sc...
research
12/01/2020

Transfer learning to enhance amenorrhea status prediction in cancer and fertility data with missing values

Collecting sufficient labelled training data for health and medical prob...
research
10/16/2019

Evolution of transfer learning in natural language processing

In this paper, we present a study of the recent advancements which have ...
research
01/14/2022

Adaptive Transfer Learning for Plant Phenotyping

Plant phenotyping (Guo et al. 2021; Pieruschka et al. 2019) focuses on s...
research
09/10/2021

Investigating Numeracy Learning Ability of a Text-to-Text Transfer Model

The transformer-based pre-trained language models have been tremendously...
research
02/27/2020

A Primer in BERTology: What we know about how BERT works

Transformer-based models are now widely used in NLP, but we still do not...
research
02/21/2022

BERT WEAVER: Using WEight AVERaging to Enable Lifelong Learning for Transformer-based Models

Recent developments in transfer learning have boosted the advancements i...

Please sign up or login with your details

Forgot password? Click here to reset