Enlightening Deep Neural Networks with Knowledge of Confounding Factors

07/08/2016
by   Yu Zhong, et al.
0

Deep learning techniques have demonstrated significant capacity in modeling some of the most challenging real world problems of high complexity. Despite the popularity of deep models, we still strive to better understand the underlying mechanism that drives their success. Motivated by observations that neurons in trained deep nets predict attributes indirectly related to the training tasks, we recognize that a deep network learns representations more general than the task at hand to disentangle impacts of multiple confounding factors governing the data, in order to isolate the effects of the concerning factors and optimize a given objective. Consequently, we propose a general framework to augment training of deep models with information on auxiliary explanatory data variables, in an effort to boost this disentanglement and train deep networks that comprehend the data interactions and distributions more accurately, and thus improve their generalizability. We incorporate information on prominent auxiliary explanatory factors of the data population into existing architectures as secondary objective/loss blocks that take inputs from hidden layers during training. Once trained, these secondary circuits can be removed to leave a model with the same architecture as the original, but more generalizable and discerning thanks to its comprehension of data interactions. Since pose is one of the most dominant confounding factors for object recognition, we apply this principle to instantiate a pose-aware deep convolutional neural network and demonstrate that auxiliary pose information indeed improves the classification accuracy in our experiments on SAR target classification tasks.

READ FULL TEXT

page 5

page 6

page 8

research
03/03/2019

Accelerating Training of Deep Neural Networks with a Standardization Loss

A significant advance in accelerating neural network training has been t...
research
09/18/2017

Wide and deep volumetric residual networks for volumetric image classification

3D shape models that directly classify objects from 3D information have ...
research
10/31/2018

Deep Generative Model with Beta Bernoulli Process for Modeling and Learning Confounding Factors

While deep representation learning has become increasingly capable of se...
research
12/12/2018

Bridging the Generalization Gap: Training Robust Models on Confounded Biological Data

Statistical learning on biological data can be challenging due to confou...
research
07/05/2017

Data-Driven Sparse Structure Selection for Deep Neural Networks

Deep convolutional neural networks have liberated its extraordinary powe...
research
01/19/2019

Overfitting Mechanism and Avoidance in Deep Neural Networks

Assisted by the availability of data and high performance computing, dee...
research
12/10/2019

Deep Relevance Regularization: Interpretable and Robust Tumor Typing of Imaging Mass Spectrometry Data

Neural networks have recently been established as a viable classificatio...

Please sign up or login with your details

Forgot password? Click here to reset