🤖 AI Summary
In particle physics, composite hypothesis testing suffers from uncontrolled statistical power distribution: the generalized likelihood ratio test (LRT) lacks theoretical optimality, and conventional methods cannot actively concentrate power in physically relevant parameter regions. This work proposes a tunable test statistic that explicitly shapes power across the parameter space via a physics-motivated weighting function; combined with a machine-learning-accelerated Neyman construction, it ensures strict coverage of confidence intervals. Evaluated on ATLAS Higgs→ττ simulated data and LZ dark matter search scenarios, the method significantly enhances signal detection sensitivity and parameter estimation accuracy. Relative to the standard LRT, it achieves an average 20–35% improvement in statistical power within targeted parameter regions. This constitutes the first optimization framework for composite testing in high-energy physics that simultaneously satisfies theoretical rigor—guaranteeing coverage—and practical utility—enabling physics-driven power allocation.
📝 Abstract
Particle physics experiments rely on the (generalised) likelihood ratio test (LRT) for searches and measurements, which consist of composite hypothesis tests. However, this test is not guaranteed to be optimal, as the Neyman-Pearson lemma pertains only to simple hypothesis tests. Any choice of test statistic thus implicitly determines how statistical power varies across the parameter space. An improvement in the core statistical testing methodology for general settings with composite tests would have widespread ramifications across experiments. We discuss an alternate test statistic that provides the data analyzer an ability to focus the power of the test on physics-motivated regions of the parameter space. We demonstrate the improvement from this technique compared to the LRT on a Higgs $
ightarrowττ$ dataset simulated by the ATLAS experiment and a dark matter dataset inspired by the LZ experiment. We also employ machine learning to efficiently perform the Neyman construction, which is essential to ensure statistically valid confidence intervals.