🤖 AI Summary
High-dimensional Bayesian optimization (BO) suffers from the “curse of dimensionality,” with existing structure-exploiting approaches—e.g., those assuming sparsity or smoothness—often underperforming even Bayesian linear regression, revealing fundamental limitations of conventional heuristics. We identify the root cause as boundary-seeking behavior of standard kernel functions in high dimensions. To address this, we propose a minimalist yet effective paradigm: Gaussian processes with linear kernels, augmented by invertible geometric transformations to correct distributional shift and closed-form sampling for linear-time computational scaling. Our method consistently outperforms state-of-the-art high-dimensional BO algorithms across 60–6,000 dimensions and achieves new SOTA performance on ultra-large-scale tasks such as molecular optimization involving over 20,000 function evaluations. Crucially, it dispenses with complex structural priors, establishing a simpler, more scalable design principle for high-dimensional BO.
📝 Abstract
High-dimensional spaces have challenged Bayesian optimization (BO). Existing methods aim to overcome this so-called curse of dimensionality by carefully encoding structural assumptions, from locality to sparsity to smoothness, into the optimization procedure. Surprisingly, we demonstrate that these approaches are outperformed by arguably the simplest method imaginable: Bayesian linear regression. After applying a geometric transformation to avoid boundary-seeking behavior, Gaussian processes with linear kernels match state-of-the-art performance on tasks with 60- to 6,000-dimensional search spaces. Linear models offer numerous advantages over their non-parametric counterparts: they afford closed-form sampling and their computation scales linearly with data, a fact we exploit on molecular optimization tasks with > 20,000 observations. Coupled with empirical analyses, our results suggest the need to depart from past intuitions about BO methods in high-dimensional spaces.