TaskWeb: Selecting Better Source Tasks for Multi-task NLP

05/22/2023
by   Joongwon Kim, et al.
0

Recent work in NLP has shown promising results in training models on large amounts of tasks to achieve better generalization. However, it is not well-understood how tasks are related, and how helpful training tasks can be chosen for a new task. In this work, we investigate whether knowing task relationships via pairwise task transfer improves choosing one or more source tasks that help to learn a new target task. We provide TaskWeb, a large-scale benchmark of pairwise task transfers for 22 NLP tasks using three different model types, sizes, and adaptation methods, spanning about 25,000 experiments. Then, we design a new method TaskShop based on our analysis of TaskWeb. TaskShop uses TaskWeb to estimate the benefit of using a source task for learning a new target, and to choose a subset of helpful training tasks for multi-task learning. Our method improves overall rankings and top-k precision of source tasks by 12 smaller multi-task training sets that improve zero-shot performances across 11 different target tasks by at least 4.3

READ FULL TEXT

page 4

page 17

page 18

page 19

page 20

page 21

research
04/07/2022

A Survey of Multi-task Learning in Natural Language Processing: Regarding Task Relatedness and Training Methods

Multi-task learning (MTL) has become increasingly popular in natural lan...
research
04/23/2022

Exploring the Role of Task Transferability in Large-Scale Multi-Task Learning

Recent work has found that multi-task training with a large number of di...
research
05/17/2022

When to Use Multi-Task Learning vs Intermediate Fine-Tuning for Pre-Trained Encoder Transfer Learning

Transfer learning (TL) in natural language processing (NLP) has seen a s...
research
11/05/2016

A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks

Transfer and multi-task learning have traditionally focused on either a ...
research
03/21/2018

Olive Oil is Made of Olives, Baby Oil is Made for Babies: Interpreting Noun Compounds using Paraphrases in a Neural Model

Automatic interpretation of the relation between the constituents of a n...
research
03/25/2023

Identification of Negative Transfers in Multitask Learning Using Surrogate Models

Multitask learning is widely used in practice to train a low-resource ta...
research
05/31/2022

Multi-task Optimization Based Co-training for Electricity Consumption Prediction

Real-world electricity consumption prediction may involve different task...

Please sign up or login with your details

Forgot password? Click here to reset