Learn Faster and Forget Slower via Fast and Stable Task Adaptation

07/02/2020
by   Farshid Varno, et al.
0

Training Deep Neural Networks (DNNs) is still highly time-consuming and compute-intensive. It has been shown that adapting a pretrained model may significantly accelerate this process. With a focus on classification, we show that current fine-tuning techniques make the pretrained models catastrophically forget the transferred knowledge even before anything about the new task is learned. Such rapid knowledge loss undermines the merits of transfer learning and may result in a much slower convergence rate compared to when the maximum amount of knowledge is exploited. We investigate the source of this problem from different perspectives and to alleviate it, introduce Fast And Stable Task-adaptation (FAST), an easy to apply fine-tuning algorithm. The paper provides a novel geometric perspective on how the loss landscape of source and target tasks are linked in different transfer learning strategies. We empirically show that compared to prevailing fine-tuning practices, FAST learns the target task faster and forgets the source task slower. The code is available at https://github.com/fvarno/FAST.

READ FULL TEXT
research
09/09/2017

Optimal Transport for Deep Joint Transfer Learning

Training a Deep Neural Network (DNN) from scratch requires a large amoun...
research
03/02/2021

Fast Adaptation with Linearized Neural Networks

The inductive biases of trained neural networks are difficult to underst...
research
06/15/2021

Learning Stable Classifiers by Transferring Unstable Features

We study transfer learning in the presence of spurious correlations. We ...
research
03/19/2023

Partial Network Cloning

In this paper, we study a novel task that enables partial knowledge tran...
research
03/05/2022

Ensemble Knowledge Guided Sub-network Search and Fine-tuning for Filter Pruning

Conventional NAS-based pruning algorithms aim to find the sub-network wi...
research
05/22/2022

muNet: Evolving Pretrained Deep Neural Networks into Scalable Auto-tuning Multitask Systems

Most uses of machine learning today involve training a model from scratc...
research
08/27/2020

A Flexible Selection Scheme for Minimum-Effort Transfer Learning

Fine-tuning is a popular way of exploiting knowledge contained in a pre-...

Please sign up or login with your details

Forgot password? Click here to reset