🤖 AI Summary
Expensive global optimization under decoupled black-box constraints poses significant computational challenges due to high evaluation costs and redundant constraint assessments.
Method: This paper proposes a novel Knowledge Gradient (KG)-based Bayesian optimization framework that dynamically identifies and prioritizes *active* (binding) constraints at the optimum. It integrates constraint relevance modeling directly into the KG acquisition function, leveraging Gaussian process surrogates to quantify each constraint’s marginal impact on objective optimality—enabling online inference of constraint importance.
Contribution/Results: The method substantially reduces unnecessary constraint evaluations while maintaining solution fidelity. In comprehensive benchmark experiments, it achieves up to 37% faster convergence and reduces constraint evaluations by over 50% compared to state-of-the-art approaches, striking an effective balance between optimization efficiency and computational economy.
📝 Abstract
Bayesian optimisation has proven to be a powerful tool for expensive global black-box optimisation problems. In this paper, we propose new Bayesian optimisation variants of the popular Knowledge Gradient acquisition functions for problems with emph{decoupled} black-box constraints, in which subsets of the objective and constraint functions may be evaluated independently. In particular, our methods aim to take into account that often only a handful of the constraints may be binding at the optimum, and hence we should evaluate only relevant constraints when trying to optimise a function. We empirically benchmark these methods against existing methods and demonstrate their superiority over the state-of-the-art.