A Few More Examples May Be Worth Billions of Parameters

10/08/2021
by   Yuval Kirstain, et al.
0

We investigate the dynamics of increasing the number of model parameters versus the number of labeled examples across a wide variety of tasks. Our exploration reveals that while scaling parameters consistently yields performance improvements, the contribution of additional examples highly depends on the task's format. Specifically, in open question answering tasks, enlarging the training set does not improve performance. In contrast, classification, extractive question answering, and multiple choice tasks benefit so much from additional examples that collecting a few hundred examples is often "worth" billions of parameters. We hypothesize that unlike open question answering, which involves recalling specific information, solving strategies for tasks with a more restricted output space transfer across examples, and can therefore be learned with small amounts of labeled data.

READ FULL TEXT

page 1

page 4

page 5

page 6

research
08/11/2021

Mounting Video Metadata on Transformer-based Language Model for Open-ended Video Question Answering

Video question answering has recently received a lot of attention from m...
research
06/07/2023

Enhancing In-Context Learning with Answer Feedback for Multi-Span Question Answering

Whereas the recent emergence of large language models (LLMs) like ChatGP...
research
09/25/2019

Question Answering is a Format; When is it Useful?

Recent years have seen a dramatic expansion of tasks and datasets posed ...
research
07/02/2020

Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering

Generative models for open domain question answering have proven to be c...
research
04/09/2020

Natural Perturbation for Robust Question Answering

While recent models have achieved human-level scores on many NLP dataset...
research
05/09/2023

MAUPQA: Massive Automatically-created Polish Question Answering Dataset

Recently, open-domain question answering systems have begun to rely heav...
research
03/05/2020

Talking-Heads Attention

We introduce "talking-heads attention" - a variation on multi-head atten...

Please sign up or login with your details

Forgot password? Click here to reset