🤖 AI Summary
This paper investigates the fundamental statistical limits of hypothesis testing in high dimensions under computational constraints, specifically focusing on the spiked Wigner model and the optimal Type-I/Type-II error trade-off achievable by polynomial-time tests.
Method: Addressing the intractability of the likelihood ratio test in high dimensions, the authors develop a novel hardness analysis framework integrating low-degree moment analysis with achievability proofs. They leverage the low-degree likelihood ratio, the strengthened low-degree hardness hypothesis, and information-computation trade-off theory.
Contribution/Results: The paper establishes the first precise asymptotic characterization of testing error under computational constraints; proves that spectral statistics—particularly linear spectral functions—are sufficient for all polynomial-time tests; and rigorously demonstrates that standard spectral methods achieve the optimal error curve among all polynomial-time tests under plausible complexity assumptions. This bridges the theoretical gap between statistical optimality and computational feasibility.
📝 Abstract
We revisit the fundamental question of simple-versus-simple hypothesis testing with an eye towards computational complexity, as the statistically optimal likelihood ratio test is often computationally intractable in high-dimensional settings. In the classical spiked Wigner model with a general i.i.d. spike prior we show (conditional on a conjecture) that an existing test based on linear spectral statistics achieves the best possible tradeoff curve between type I and type II error rates among all computationally efficient tests, even though there are exponential-time tests that do better. This result is conditional on an appropriate complexity-theoretic conjecture, namely a natural strengthening of the well-established low-degree conjecture. Our result shows that the spectrum is a sufficient statistic for computationally bounded tests (but not for all tests). To our knowledge, our approach gives the first tool for reasoning about the precise asymptotic testing error achievable with efficient computation. The main ingredients required for our hardness result are a sharp bound on the norm of the low-degree likelihood ratio along with (counterintuitively) a positive result on achievability of testing. This strategy appears to be new even in the setting of unbounded computation, in which case it gives an alternate way to analyze the fundamental statistical limits of testing.