Towards Robust and Efficient Continual Language Learning

07/11/2023
by   Adam Fisch, et al.
0

As the application space of language models continues to evolve, a natural question to ask is how we can quickly adapt models to new tasks. We approach this classic question from a continual learning perspective, in which we aim to continue fine-tuning models trained on past tasks on new tasks, with the goal of "transferring" relevant knowledge. However, this strategy also runs the risk of doing more harm than good, i.e., negative transfer. In this paper, we construct a new benchmark of task sequences that target different possible transfer scenarios one might face, such as a sequence of tasks with high potential of positive transfer, high potential for negative transfer, no expected effect, or a mixture of each. An ideal learner should be able to maximally exploit information from all tasks that have any potential for positive transfer, while also avoiding the negative effects of any distracting tasks that may confuse it. We then propose a simple, yet effective, learner that satisfies many of our desiderata simply by leveraging a selective strategy for initializing new models from past task checkpoints. Still, limitations remain, and we hope this benchmark can help the community to further build and analyze such learners.

READ FULL TEXT

page 5

page 15

page 16

page 18

research
04/04/2023

I2I: Initializing Adapters with Improvised Knowledge

Adapters present a promising solution to the catastrophic forgetting pro...
research
03/21/2023

Continual Learning in the Presence of Spurious Correlation

Most continual learning (CL) algorithms have focused on tackling the sta...
research
10/10/2015

Attend, Adapt and Transfer: Attentive Deep Architecture for Adaptive Transfer from multiple sources in the same domain

Transferring knowledge from prior source tasks in solving a new target t...
research
05/24/2022

Continual-T0: Progressively Instructing 50+ Tasks to Language Models Without Forgetting

Recent work on large language models relies on the intuition that most n...
research
12/18/2021

Continual Learning with Knowledge Transfer for Sentiment Classification

This paper studies continual learning (CL) for sentiment classification ...
research
02/25/2021

On continual single index learning

In this paper, we generalize the problem of single index model to the co...

Please sign up or login with your details

Forgot password? Click here to reset