Memorization Capacity of Neural Networks with Conditional Computation

03/20/2023
by   Erdem Koyuncu, et al.
0

Many empirical studies have demonstrated the performance benefits of conditional computation in neural networks, including reduced inference time and power consumption. We study the fundamental limits of neural conditional computation from the perspective of memorization capacity. For Rectified Linear Unit (ReLU) networks without conditional computation, it is known that memorizing a collection of n input-output relationships can be accomplished via a neural network with O(√(n)) neurons. Calculating the output of this neural network can be accomplished using O(√(n)) elementary arithmetic operations of additions, multiplications and comparisons for each input. Using a conditional ReLU network, we show that the same task can be accomplished using only O(log n) operations per input. This represents an almost exponential improvement as compared to networks without conditional computation. We also show that the Θ(log n) rate is the best possible. Our achievability result utilizes a general methodology to synthesize a conditional network out of an unconditional network in a computationally-efficient manner, bridging the gap between unconditional and conditional architectures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/01/2021

Computation complexity of deep ReLU neural networks in high-dimensional approximation

The purpose of the present paper is to study the computation complexity ...
research
11/20/2020

Computation capacities of a broad class of signaling networks are higher than their communication capacities

Due to structural and functional abnormalities or genetic variations and...
research
02/06/2019

On the CVP for the root lattices via folding with deep ReLU neural networks

Point lattices and their decoding via neural networks are considered in ...
research
01/22/2016

Bitwise Neural Networks

Based on the assumption that there exists a neural network that efficien...
research
12/03/2022

Probabilistic Verification of ReLU Neural Networks via Characteristic Functions

Verifying the input-output relationships of a neural network so as to ac...
research
02/17/2020

Controlling Computation versus Quality for Neural Sequence Models

Most neural networks utilize the same amount of compute for every exampl...
research
11/24/2015

Dynamic Capacity Networks

We introduce the Dynamic Capacity Network (DCN), a neural network that c...

Please sign up or login with your details

Forgot password? Click here to reset