Representation Benefits of Deep Feedforward Networks

09/27/2015
by   Matus Telgarsky, et al.
0

This note provides a family of classification problems, indexed by a positive integer k, where all shallow networks with fewer than exponentially (in k) many nodes exhibit error at least 1/6, whereas a deep network with 2 nodes in each of 2k layers achieves zero error, as does a recurrent network with 3 distinct nodes iterated k times. The proof is elementary, and the networks are standard feedforward networks with ReLU (Rectified Linear Unit) nonlinearities.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset