Optimal Transport for Deep Joint Transfer Learning

09/09/2017
by   Ying Lu, et al.
0

Training a Deep Neural Network (DNN) from scratch requires a large amount of labeled data. For a classification task where only small amount of training data is available, a common solution is to perform fine-tuning on a DNN which is pre-trained with related source data. This consecutive training process is time consuming and does not consider explicitly the relatedness between different source and target tasks. In this paper, we propose a novel method to jointly fine-tune a Deep Neural Network with source data and target data. By adding an Optimal Transport loss (OT loss) between source and target classifier predictions as a constraint on the source classifier, the proposed Joint Transfer Learning Network (JTLN) can effectively learn useful knowledge for target classification from source data. Furthermore, by using different kind of metric as cost matrix for the OT loss, JTLN can incorporate different prior knowledge about the relatedness between target categories and source categories. We carried out experiments with JTLN based on Alexnet on image classification datasets and the results verify the effectiveness of the proposed JTLN in comparison with standard consecutive fine-tuning. This Joint Transfer Learning with OT loss is general and can also be applied to other kind of Neural Networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/21/2019

AdaFilter: Adaptive Filter Fine-tuning for Deep Transfer Learning

There is an increasing number of pre-trained deep neural network models....
research
07/02/2020

Learn Faster and Forget Slower via Fast and Stable Task Adaptation

Training Deep Neural Networks (DNNs) is still highly time-consuming and ...
research
02/28/2017

Borrowing Treasures from the Wealthy: Deep Transfer Learning through Selective Joint Fine-tuning

Deep neural networks require a large amount of labeled training data dur...
research
09/13/2019

Semantic and Visual Similarities for Efficient Knowledge Transfer in CNN Training

In recent years, representation learning approaches have disrupted many ...
research
02/02/2022

Auto-Transfer: Learning to Route Transferrable Representations

Knowledge transfer between heterogeneous source and target networks and ...
research
11/29/2022

Transfer Learning with Uncertainty Quantification: Random Effect Calibration of Source to Target (RECaST)

Transfer learning uses a data model, trained to make predictions or infe...
research
11/01/2020

An Information-Geometric Distance on the Space of Tasks

This paper computes a distance between tasks modeled as joint distributi...

Please sign up or login with your details

Forgot password? Click here to reset