SOTAVerified

Recovering the Lowest Layer of Deep Networks with High Threshold Activations

2019-03-21ICLR 2019Unverified0· sign in to hype

Surbhi Goel, Rina Panigrahy

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Giving provable guarantees for learning neural networks is a core challenge of machine learning theory. Most prior work gives parameter recovery guarantees for one hidden layer networks, however, the networks used in practice have multiple non-linear layers. In this work, we show how we can strengthen such results to deeper networks -- we address the problem of uncovering the lowest layer in a deep neural network under the assumption that the lowest layer uses a high threshold before applying the activation, the upper network can be modeled as a well-behaved polynomial and the input distribution is Gaussian.

Tasks

Reproductions