SOTAVerified

Learning and Testing Junta Distributions with Subcube Conditioning

2020-04-26Unverified0· sign in to hype

Xi Chen, Rajesh Jayaram, Amit Levi, Erik Waingarten

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We study the problems of learning and testing junta distributions on \-1,1\^n with respect to the uniform distribution, where a distribution p is a k-junta if its probability mass function p(x) depends on a subset of at most k variables. The main contribution is an algorithm for finding relevant coordinates in a k-junta distribution with subcube conditioning [BC18, CCKLW20]. We give two applications: 1. An algorithm for learning k-junta distributions with O(k/^2) n + O(2^k/^2) subcube conditioning queries, and 2. An algorithm for testing k-junta distributions with O((k + n)/^2) subcube conditioning queries. All our algorithms are optimal up to poly-logarithmic factors. Our results show that subcube conditioning, as a natural model for accessing high-dimensional distributions, enables significant savings in learning and testing junta distributions compared to the standard sampling model. This addresses an open question posed by Aliakbarpour, Blais, and Rubinfeld [ABR17].

Tasks

Reproductions