🤖 AI Summary
High-dimensional Bayesian optimization (BO) suffers from data sparsity and poor scalability of surrogate models. While existing variational autoencoder (VAE)-based approaches mitigate this via dimensionality reduction, their reconstruction objective induces distributional mismatch between the latent and original spaces—degrading optimization performance. This paper proposes HiBBO, the first BO framework integrating the HiPPO (Hierarchical Permutation-invariant Projection Operator) sequence modeling mechanism into the VAE architecture. HiPPO explicitly enforces consistency of function responses in the latent space, thereby improving distributional fidelity. By preserving both spatial and statistical properties of high-dimensional functions within a low-dimensional latent representation, HiBBO accelerates convergence and enhances solution quality. Experiments demonstrate that HiBBO outperforms state-of-the-art VAE-BO methods across multiple high-dimensional benchmarks and exhibits strong generalization in real-world applications—including neural architecture search and materials science.
📝 Abstract
Bayesian Optimisation (BO) is a powerful tool for optimising expensive blackbox functions but its effectiveness diminishes in highdimensional spaces due to sparse data and poor surrogate model scalability While Variational Autoencoder (VAE) based approaches address this by learning low-dimensional latent representations the reconstructionbased objective function often brings the functional distribution mismatch between the latent space and original space leading to suboptimal optimisation performance In this paper we first analyse the reason why reconstructiononly loss may lead to distribution mismatch and then propose HiBBO a novel BO framework that introduces the space consistency into the latent space construction in VAE using HiPPO - a method for longterm sequence modelling - to reduce the functional distribution mismatch between the latent space and original space Experiments on highdimensional benchmark tasks demonstrate that HiBBO outperforms existing VAEBO methods in convergence speed and solution quality Our work bridges the gap between high-dimensional sequence representation learning and efficient Bayesian Optimisation enabling broader applications in neural architecture search materials science and beyond.