How transferable are features in deep neural networks?

11/06/2014
by   Jason Yosinski, et al.
0

Many deep neural networks trained on natural images exhibit a curious phenomenon in common: on the first layer they learn features similar to Gabor filters and color blobs. Such first-layer features appear not to be specific to a particular dataset or task, but general in that they are applicable to many datasets and tasks. Features must eventually transition from general to specific by the last layer of the network, but this transition has not been studied extensively. In this paper we experimentally quantify the generality versus specificity of neurons in each layer of a deep convolutional neural network and report a few surprising results. Transferability is negatively affected by two distinct issues: (1) the specialization of higher layer neurons to their original task at the expense of performance on the target task, which was expected, and (2) optimization difficulties related to splitting networks between co-adapted neurons, which was not expected. In an example network trained on ImageNet, we demonstrate that either of these two issues may dominate, depending on whether features are transferred from the bottom, middle, or top of the network. We also document that the transferability of features decreases as the distance between the base task and target task increases, but that transferring features even from distant tasks can be better than using random features. A final surprising result is that initializing a network with transferred features from almost any number of layers can produce a boost to generalization that lingers even after fine-tuning to the target dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/26/2019

Towards Understanding the Transferability of Deep Representations

Deep neural networks trained on a wide range of datasets demonstrate imp...
research
09/25/2019

Wider Networks Learn Better Features

Transferability of learned features between tasks can massively reduce t...
research
04/29/2019

ConvTimeNet: A Pre-trained Deep Convolutional Neural Network for Time Series Classification

Training deep neural networks often requires careful hyper-parameter tun...
research
04/20/2018

CactusNets: Layer Applicability as a Metric for Transfer Learning

Deep neural networks trained over large datasets learn features that are...
research
12/01/2022

Rethinking Two Consensuses of the Transferability in Deep Learning

Deep transfer learning (DTL) has formed a long-term quest toward enablin...
research
05/20/2017

Forward Thinking: Building Deep Random Forests

The success of deep neural networks has inspired many to wonder whether ...
research
04/23/2018

Parameter Transfer Unit for Deep Neural Networks

Parameters in deep neural networks which are trained on large-scale data...

Please sign up or login with your details

Forgot password? Click here to reset