GradTS: A Gradient-Based Automatic Auxiliary Task Selection Method Based on Transformer Networks

09/13/2021
by   Weicheng Ma, et al.
0

A key problem in multi-task learning (MTL) research is how to select high-quality auxiliary tasks automatically. This paper presents GradTS, an automatic auxiliary task selection method based on gradient calculation in Transformer-based models. Compared to AUTOSEM, a strong baseline method, GradTS improves the performance of MT-DNN with a bert-base-cased backend model, from 0.33 benchmarks. GradTS is also time-saving since (1) its gradient calculations are based on single-task experiments and (2) the gradients are re-used without additional experiments when the candidate task set changes. On the 8 GLUE classification tasks, for example, GradTS costs on average 21.32 than AUTOSEM with comparable GPU consumption. Further, we show the robustness of GradTS across various task settings and model selections, e.g. mixed objectives among candidate tasks. The efficiency and efficacy of GradTS in these case studies illustrate its general applicability in MTL research without requiring manual task filtering or costly parameter tuning.

READ FULL TEXT

page 4

page 12

research
04/08/2019

AutoSeM: Automatic Task Selection and Mixing in Multi-Task Learning

Multi-task learning (MTL) has achieved success over a wide range of prob...
research
08/10/2022

Multi-task Active Learning for Pre-trained Transformer-based Models

Multi-task learning, in which several tasks are jointly learned by a sin...
research
03/14/2022

MetaBalance: Improving Multi-Task Recommendations via Adapting Gradient Magnitudes of Auxiliary Tasks

In many personalized recommendation scenarios, the generalization abilit...
research
04/09/2022

Efficient Extraction of Pathologies from C-Spine Radiology Reports using Multi-Task Learning

Pretrained Transformer based models finetuned on domain specific corpora...
research
10/08/2021

ALL-IN-ONE: Multi-Task Learning BERT models for Evaluating Peer Assessments

Peer assessment has been widely applied across diverse academic fields o...
research
07/12/2021

A Flexible Multi-Task Model for BERT Serving

In this demonstration, we present an efficient BERT-based multi-task (MT...
research
07/17/2017

Auxiliary Objectives for Neural Error Detection Models

We investigate the utility of different auxiliary objectives and trainin...

Please sign up or login with your details

Forgot password? Click here to reset