🤖 AI Summary
In quantum machine learning, angle encoding induces hybrid frequencies that grow doubly exponentially, causing parameter explosion; even when target functions possess critical spectral components, models fail to train due to insufficient parameters. This work identifies and alleviates this parameter trainability bottleneck via a novel frequency selection and dimension separation strategy. Specifically, we perform frequency-domain analysis to identify an effective subset of spectral components and enforce dimension-wise decoupling to curb parameter growth. Our method builds upon angle-encoded feature maps and integrates white-box function verification with parameter-constrained optimization. It enables end-to-end training and inference on noisy quantum simulators and real hardware (IBM Quantum). Experiments demonstrate successful fitting of two classes of high-dimensional functions that conventional QML approaches fail to converge on. While preserving accuracy, our approach reduces parameter counts by one to two orders of magnitude, substantially expanding the problem scale and dimensionality feasible on current NISQ devices.
📝 Abstract
To leverage the potential computational speedup of quantum computing (QC), research in quantum machine learning (QML) has gained increasing prominence. Angle encoding techniques in QML models have been shown to generate truncated Fourier series, offering asymptotically universal function approximation capabilities. By selecting efficient feature maps (FMs) within quantum circuits, one can leverage the exponential growth of Fourier frequencies for improved approximation. In multi-dimensional settings, additional input dimensions induce further exponential scaling via mixed frequencies. In practice, however, quantum models frequently fail at regression tasks. Through two white-box experiments, we show that such failures can occur even when the relevant frequencies are present, due to an insufficient number of trainable parameters.
In order to mitigate the double-exponential parameter growth resulting from double-exponentially growing frequencies, we propose frequency selection and dimensional separation as techniques to constrain the number of parameters, thereby improving trainability. By restricting the QML model to essential frequencies and permitting mixed frequencies only among feature dimensions with known interdependence, we expand the set of tractable problems on current hardware. We demonstrate the reduced parameter requirements by fitting two white-box functions with known frequency spectrum and dimensional interdependencies that could not be fitted with the default methods. The reduced parameter requirements permit us to perform training on a noisy quantum simulator and to demonstrate inference on real quantum hardware.