SOTAVerified

Efficiently Learning One Hidden Layer ReLU Networks From Queries

2021-12-01NeurIPS 2021Unverified0· sign in to hype

Sitan Chen, Adam Klivans, Raghu Meka

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

While the problem of PAC learning neural networks from samples has received considerable attention in recent years, in certain settings like model extraction attacks, it is reasonable to imagine having more than just the ability to observe random labeled examples. Motivated by this, we consider the following problem: given black-box query access to a neural network F, recover F up to some error. Formally, we show that if F is an arbitrary one hidden layer neural network with ReLU activations, there is an algorithm with query complexity and runtime polynomial in all parameters which outputs a network F’ achieving low square loss relative to F with respect to the Gaussian measure. While a number of works in the security literature have proposed and empirically demonstrated the effectiveness of certain algorithms for this problem, ours is to the best of our knowledge the first provable guarantee in this vein.

Tasks

Reproductions