Dynamic Normalization

01/15/2021
by   Chuan Liu, et al.
0

Batch Normalization has become one of the essential components in CNN. It allows the network to use a higher learning rate and speed up training. And the network doesn't need to be initialized carefully. However, in our work, we find that a simple extension of BN can increase the performance of the network. First, we extend BN to adaptively generate scale and shift parameters for each mini-batch data, called DN-C (Batch-shared and Channel-wise). We use the statistical characteristics of mini-batch data (E[X], Std[X]∈ℝ^c) as the input of SC module. Then we extend BN to adaptively generate scale and shift parameters for each channel of each sample, called DN-B (Batch and Channel-wise). Our experiments show that DN-C model can't train normally, but DN-B model has very good robustness. In classification task, DN-B can improve the accuracy of the MobileNetV2 on ImageNet-100 more than 2 task, DN-B can improve the accuracy of the SSDLite on MS-COCO nearly 4 with the same settings. Compared with BN, DN-B has stable performance when using higher learning rate or smaller batch size.

READ FULL TEXT

page 6

page 7

research
11/20/2017

MegDet: A Large Mini-Batch Object Detector

The improvements in recent CNN-based object detection works, from R-CNN ...
research
02/13/2020

Scalable and Practical Natural Gradient for Large-Scale Deep Learning

Large-scale distributed training of deep neural networks results in mode...
research
03/22/2021

Delving into Variance Transmission and Normalization: Shift of Average Gradient Makes the Network Collapse

Normalization operations are essential for state-of-the-art neural netwo...
research
06/07/2016

Systematic evaluation of CNN advances on the ImageNet

The paper systematically studies the impact of a range of recent advance...
research
08/04/2019

Attentive Normalization

Batch Normalization (BN) is a vital pillar in the development of deep le...
research
04/23/2020

YOLOv4: Optimal Speed and Accuracy of Object Detection

There are a huge number of features which are said to improve Convolutio...
research
06/08/2020

Passive Batch Injection Training Technique: Boosting Network Performance by Injecting Mini-Batches from a different Data Distribution

This work presents a novel training technique for deep neural networks t...

Please sign up or login with your details

Forgot password? Click here to reset