Representation Benefits of Deep Feedforward Networks
2015-09-27Unverified0· sign in to hype
Matus Telgarsky
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
This note provides a family of classification problems, indexed by a positive integer k, where all shallow networks with fewer than exponentially (in k) many nodes exhibit error at least 1/6, whereas a deep network with 2 nodes in each of 2k layers achieves zero error, as does a recurrent network with 3 distinct nodes iterated k times. The proof is elementary, and the networks are standard feedforward networks with ReLU (Rectified Linear Unit) nonlinearities.