๐ค AI Summary
This paper addresses the first rigorous theoretical analysis of approximation error in Structured Kernel Interpolation (SKI) for high-dimensional Gaussian processes (GPs), focusing on the fundamental trade-off between approximation accuracy and linear-time computational efficiency as dimensionality increases.
Method: We derive a spectral norm error bound for the SKI Gram matrix, integrating asymptotic complexity analysis with the convolutional cubic interpolation structure to uncover an intrinsic dichotomy in error-efficiency compatibility between low dimensions (d โค 3) and higher dimensions (d > 3).
Contributions/Results: (1) We provide a tight theoretical characterization of SKI approximation error; (2) we prove that arbitrary-accuracy O(n) inference is achievable for d โค 3โestablishing the first precise, controllable condition for linear-time GPs; and (3) we derive the optimal inducing point count n^{d/3}, enabling principled joint optimization of hyperparameters and computational resources.
๐ Abstract
Structured Kernel Interpolation (SKI) (Wilson et al. 2015) helps scale Gaussian Processes (GPs) by approximating the kernel matrix via interpolation at inducing points, achieving linear computational complexity. However, it lacks rigorous theoretical error analysis. This paper bridges the gap: we prove error bounds for the SKI Gram matrix and examine the error's effect on hyperparameter estimation and posterior inference. We further provide a practical guide to selecting the number of inducing points under convolutional cubic interpolation: they should grow as $n^{d/3}$ for error control. Crucially, we identify two dimensionality regimes governing the trade-off between SKI Gram matrix spectral norm error and computational complexity. For $d leq 3$, any error tolerance can achieve linear time for sufficiently large sample size. For $d>3$, the error must increase with sample size to maintain linear time. Our analysis provides key insights into SKI's scalability-accuracy trade-offs, establishing precise conditions for achieving linear-time GP inference with controlled approximation error.