Discriminability-enforcing loss to improve representation learning

02/14/2022
by   Florinel-Alin Croitoru, et al.
0

During the training process, deep neural networks implicitly learn to represent the input data samples through a hierarchy of features, where the size of the hierarchy is determined by the number of layers. In this paper, we focus on enforcing the discriminative power of the high-level representations, that are typically learned by the deeper layers (closer to the output). To this end, we introduce a new loss term inspired by the Gini impurity, which is aimed at minimizing the entropy (increasing the discriminative power) of individual high-level features with respect to the class labels. Although our Gini loss induces highly-discriminative features, it does not ensure that the distribution of the high-level features matches the distribution of the classes. As such, we introduce another loss term to minimize the Kullback-Leibler divergence between the two distributions. We conduct experiments on two image classification data sets (CIFAR-100 and Caltech 101), considering multiple neural architectures ranging from convolutional networks (ResNet-17, ResNet-18, ResNet-50) to transformers (CvT). Our empirical results show that integrating our novel loss terms into the training objective consistently outperforms the models trained with cross-entropy alone.

READ FULL TEXT
research
08/07/2020

Revisiting Mid-Level Patterns for Distant-Domain Few-Shot Recognition

Cross-domain few-shot learning (FSL) is proposed recently to transfer kn...
research
03/08/2020

Inverse Feature Learning: Feature learning based on Representation Learning of Error

This paper proposes inverse feature learning as a novel supervised featu...
research
03/23/2017

Is Second-order Information Helpful for Large-scale Visual Recognition?

By stacking layers of convolution and nonlinearity, convolutional networ...
research
06/30/2019

Learning to Find Correlated Features by Maximizing Information Flow in Convolutional Neural Networks

Training convolutional neural networks for image classification tasks us...
research
07/05/2023

How Deep Neural Networks Learn Compositional Data: The Random Hierarchy Model

Learning generic high-dimensional tasks is notably hard, as it requires ...
research
12/05/2016

Local Blur Mapping: Exploiting High-Level Semantics by Deep Neural Networks

The human visual system excels at detecting local blur of visual images,...
research
04/07/2017

Jet Constituents for Deep Neural Network Based Top Quark Tagging

Recent literature on deep neural networks for tagging of highly energeti...

Please sign up or login with your details

Forgot password? Click here to reset