Old BERT, New Tricks: Artificial Language Learning for Pre-Trained Language Models

09/13/2021
by   Lisa Bylinina, et al.
0

We extend the artificial language learning experimental paradigm from psycholinguistics and apply it to pre-trained language models – specifically, BERT (Devlin et al., 2019). We treat the model as a subject in an artificial language learning experimental setting: in order to learn the relation between two linguistic properties A and B, we introduce a set of new, non-existent, linguistic items, give the model information about their variation along property A, then measure to what extent the model learns property B for these items as a result of training. We show this method at work for degree modifiers (expressions like "slightly", "very", "rather", "extremely") and test the hypothesis that the degree expressed by modifiers (low, medium or high degree) is related to their sensitivity to sentence polarity (whether they show preference for affirmative or negative sentences or neither). Our experimental results are compatible with existing linguistic observations that relate degree semantics to polarity-sensitivity, including the main one: low degree semantics leads to positive polarity sensitivity (that is, to preference towards affirmative contexts). The method can be used in linguistics to elaborate on hypotheses and interpret experimental results, as well as for more insightful evaluation of linguistic representations in language models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2021

Transformers in the loop: Polarity in neural models of language

Representation of linguistic phenomena in computational language models ...
research
06/20/2023

Towards Understanding What Code Language Models Learned

Pre-trained language models are effective in a variety of natural langua...
research
08/31/2018

Do Language Models Understand Anything? On the Ability of LSTMs to Understand Negative Polarity Items

In this paper, we attempt to link the inner workings of a neural languag...
research
07/04/2022

Probing via Prompting

Probing is a popular method to discern what linguistic information is co...
research
03/12/2021

Improving Authorship Verification using Linguistic Divergence

We propose an unsupervised solution to the Authorship Verification task ...
research
09/20/2023

Exploring the Relationship between LLM Hallucinations and Prompt Linguistic Nuances: Readability, Formality, and Concreteness

As Large Language Models (LLMs) have advanced, they have brought forth n...
research
12/23/2019

Probing the phonetic and phonological knowledge of tones in Mandarin TTS models

This study probes the phonetic and phonological knowledge of lexical ton...

Please sign up or login with your details

Forgot password? Click here to reset