π€ AI Summary
This paper studies the problem of learning $k$-junta distributions over the Boolean hypercube ${0,1}^n$, i.e., distributions whose probability mass functions depend on at most $k$ variables. It establishes, for the first time, a computational equivalence between junta distribution learning and noisy $k$-parity learning (LPN). Leveraging Fourier analysis, low-order correlation screening, and noise-robust moment estimation, the authors design a statistically optimal and computationally efficient algorithm: it achieves the tight sample complexity $O(2^k log n)$βup to polylogarithmic factorsβand matches the runtime of the best prior algorithms. Moreover, they prove that any substantial improvement in sample or time complexity would imply a breakthrough in solving LPN. This work thus settles the theoretical tight bounds for junta distribution learning and introduces a novel paradigm for learning high-dimensional sparse distributions.
π Abstract
We study the problem of learning junta distributions on ${0, 1}^n$, where a distribution is a $k$-junta if its probability mass function depends on a subset of at most $k$ variables. We make two main contributions: - We show that learning $k$-junta distributions is emph{computationally} equivalent to learning $k$-parity functions with noise (LPN), a landmark problem in computational learning theory. - We design an algorithm for learning junta distributions whose statistical complexity is optimal, up to polylogarithmic factors. Computationally, our algorithm matches the complexity of previous (non-sample-optimal) algorithms. Combined, our two contributions imply that our algorithm cannot be significantly improved, statistically or computationally, barring a breakthrough for LPN.