DNN Feature Map Compression using Learned Representation over GF(2)

08/15/2018
by   Denis A. Gudovskiy, et al.
0

In this paper, we introduce a method to compress intermediate feature maps of deep neural networks (DNNs) to decrease memory storage and bandwidth requirements during inference. Unlike previous works, the proposed method is based on converting fixed-point activations into vectors over the smallest GF(2) finite field followed by nonlinear dimensionality reduction (NDR) layers embedded into a DNN. Such an end-to-end learned representation finds more compact feature maps by exploiting quantization redundancies within the fixed-point activations along the channel or spatial dimensions. We apply the proposed network architectures derived from modified SqueezeNet and MobileNetV2 to the tasks of ImageNet classification and PASCAL VOC object detection. Compared to prior approaches, the conducted experiments show a factor of 2 decrease in memory requirements with minor degradation in accuracy while adding only bitwise computations.

READ FULL TEXT
research
03/08/2017

Deep Convolutional Neural Network Inference with Floating-point Weights and Fixed-point Activations

Deep convolutional neural network (CNN) inference requires significant a...
research
03/22/2022

FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation

Deep neural networks (DNNs) have demonstrated their effectiveness in a w...
research
11/19/2019

IFQ-Net: Integrated Fixed-point Quantization Networks for Embedded Vision

Deploying deep models on embedded devices has been a challenging problem...
research
03/15/2022

Implicit Feature Decoupling with Depthwise Quantization

Quantization has been applied to multiple domains in Deep Neural Network...
research
04/29/2021

Condensation-Net: Memory-Efficient Network Architecture with Cross-Channel Pooling Layers and Virtual Feature Maps

"Lightweight convolutional neural networks" is an important research top...
research
11/07/2016

Fixed-point Factorized Networks

In recent years, Deep Neural Networks (DNN) based methods have achieved ...
research
04/13/2020

Technical Report: NEMO DNN Quantization for Deployment Model

This technical report aims at defining a formal framework for Deep Neura...

Please sign up or login with your details

Forgot password? Click here to reset