Gated Channel Transformation for Visual Recognition

09/25/2019
by   Zongxin Yang, et al.
0

In this work, we propose a generally applicable transformation unit for visual recognition with deep convolutional neural networks. This transformation explicitly models channel relationships with explainable control variables. These variables determine the neuron behaviors of competition or cooperation, and they are jointly optimized with convolutional weights towards more accurate recognition. In Squeeze-and-Excitation (SE) Networks, the channel relationships are implicitly learned by fully connected layers, and the SE block is integrated at the block-level. We instead introduce a channel normalization layer to reduce the number of parameters and computational complexity. This lightweight layer incorporates a simple L2 normalization, enabling our transformation unit applicable to operator-level without much increase of additional parameters. Extensive experiments demonstrate the effectiveness of our unit with clear margins on many vision tasks, i.e., image classification on ImageNet, object detection and instance segmentation on COCO, video classification on Kinetics.

READ FULL TEXT
research
08/04/2019

Attentive Normalization

Batch Normalization (BN) is a vital pillar in the development of deep le...
research
09/06/2019

Linear Context Transform Block

Squeeze-and-Excitation (SE) block presents a channel attention mechanism...
research
03/03/2017

Deep Collaborative Learning for Visual Recognition

Deep neural networks are playing an important role in state-of-the-art v...
research
10/22/2021

Recurrence along Depth: Deep Convolutional Neural Networks with Recurrent Layer Aggregation

This paper introduces a concept of layer aggregation to describe how inf...
research
01/06/2019

Channel Locality Block: A Variant of Squeeze-and-Excitation

Attention mechanism is a hot spot in deep learning field. Using channel ...
research
10/29/2019

Best Practices for Convolutional Neural Networks Applied to Object Recognition in Images

This research project studies the impact of convolutional neural network...
research
06/29/2020

Explainable 3D Convolutional Neural Networks by Learning Temporal Transformations

In this paper we introduce the temporally factorized 3D convolution (3TC...

Please sign up or login with your details

Forgot password? Click here to reset