Learning Activation Functions to Improve Deep Neural Networks

12/21/2014
by   Forest Agostinelli, et al.
0

Artificial neural networks typically have a fixed, non-linear activation function at each neuron. We have designed a novel form of piecewise linear activation function that is learned independently for each neuron using gradient descent. With this adaptive activation function, we are able to improve upon deep neural network architectures composed of static rectified linear units, achieving state-of-the-art performance on CIFAR-10 (7.51 CIFAR-100 (30.83 boson decay modes.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset