DropNeuron: Simplifying the Structure of Deep Neural Networks

06/23/2016
by   Wei Pan, et al.
0

Deep learning using multi-layer neural networks (NNs) architecture manifests superb power in modern machine learning systems. The trained Deep Neural Networks (DNNs) are typically large. The question we would like to address is whether it is possible to simplify the NN during training process to achieve a reasonable performance within an acceptable computational time. We presented a novel approach of optimising a deep neural network through regularisation of net- work architecture. We proposed regularisers which support a simple mechanism of dropping neurons during a network training process. The method supports the construction of a simpler deep neural networks with compatible performance with its simplified version. As a proof of concept, we evaluate the proposed method with examples including sparse linear regression, deep autoencoder and convolutional neural network. The valuations demonstrate excellent performance. The code for this work can be found in http://www.github.com/panweihit/DropNeuron

READ FULL TEXT

page 12

page 13

page 14

page 15

research
12/16/2020

Multi-Task Learning in Diffractive Deep Neural Networks via Hardware-Software Co-design

Deep neural networks (DNNs) have substantial computational requirements,...
research
09/13/2021

Explaining Deep Learning Representations by Tracing the Training Process

We propose a novel explanation method that explains the decisions of a d...
research
06/30/2020

Deep neural networks for the evaluation and design of photonic devices

The data sciences revolution is poised to transform the way photonic sys...
research
07/03/2018

Adversarial Robustness Toolbox v0.2.2

Adversarial examples have become an indisputable threat to the security ...
research
06/07/2021

Representation mitosis in wide neural networks

Deep neural networks (DNNs) defy the classical bias-variance trade-off: ...
research
05/17/2018

DNN or k-NN: That is the Generalize vs. Memorize Question

This paper studies the relationship between the classification performed...
research
03/03/2020

Deep Learning in Memristive Nanowire Networks

Analog crossbar architectures for accelerating neural network training a...

Please sign up or login with your details

Forgot password? Click here to reset