Large Scale Multi-Task Bayesian Optimization with Large Language Models

📅 2025-03-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In large-scale multi-task Bayesian optimization (≈2000 tasks), existing approaches—such as multi-task Gaussian processes and deep kernel transfer—exhibit limited knowledge transfer efficiency and marginal performance gains. Method: This paper introduces, for the first time, large language models (LLMs) into this paradigm, proposing an LLM-driven iterative transfer framework. It fine-tunes an LLM on high-quality optimization trajectories from historical tasks to generate high-potential initial points for new tasks; integrates Bayesian optimization feedback to continuously update the LLM, establishing a closed-loop cycle: trajectory encoding → LLM-based inference → Bayesian optimization execution → data refilling. Results: Evaluated on real-world antimicrobial peptide design, the method achieves significant reductions in oracle calls using only few-shot samples, outperforming standard zero-shot Bayesian optimization in both solution quality and convergence speed—effectively unifying multi-task meta-learning with generative modeling.

Technology Category

Application Category

📝 Abstract
In multi-task Bayesian optimization, the goal is to leverage experience from optimizing existing tasks to improve the efficiency of optimizing new ones. While approaches using multi-task Gaussian processes or deep kernel transfer exist, the performance improvement is marginal when scaling to more than a moderate number of tasks. We introduce a novel approach leveraging large language models (LLMs) to learn from, and improve upon, previous optimization trajectories, scaling to approximately 2000 distinct tasks. Specifically, we propose an iterative framework in which an LLM is fine-tuned using the high quality solutions produced by BayesOpt to generate improved initializations that accelerate convergence for future optimization tasks based on previous search trajectories. We evaluate our method on two distinct domains: database query optimization and antimicrobial peptide design. Results demonstrate that our approach creates a positive feedback loop, where the LLM's generated initializations gradually improve, leading to better optimization performance. As this feedback loop continues, we find that the LLM is eventually able to generate solutions to new tasks in just a few shots that are better than the solutions produced by"from scratch"by Bayesian optimization while simultaneously requiring significantly fewer oracle calls.
Problem

Research questions and friction points this paper is trying to address.

Improve multi-task Bayesian optimization efficiency
Scale optimization to approximately 2000 tasks
Enhance new task solutions with fewer oracle calls
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages large language models for multi-task optimization
Fine-tunes LLMs using Bayesian optimization solutions
Generates improved initializations for faster convergence
🔎 Similar Papers
No similar papers found.
Yimeng Zeng
Yimeng Zeng
PhD Student, University of Pennsylvania
Machine LearningBayesian OptimizationGenerative ModelsLarge Language Models
Natalie Maus
Natalie Maus
PhD. Student, University of Pennsylvania Department of Computer and Information Science
machine learningbayesian optimizationdeep learninggenerative modelingcomputational drug design
H
Haydn Thomas Jones
Department of Computer and Information Science, University of Pennsylvania, Philadelphia, PA, USA
Jeffrey Tao
Jeffrey Tao
PhD Student, University of Pennsylvania
human-computer interactiondatabases
F
Fangping Wan
Machine Biology Group, Departments of Psychiatry and Microbiology, Institute for Biomedical Informatics, Institute for Translational Medicine and Therapeutics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, USA; Departments of Bioengineering and Chemical and Biomolecular Engineering, School of Engineering and Applied Science, University of Pennsylvania, Philadelphia, PA, USA; Department of Chemistry, School of Arts and Sciences, University of Pennsylvania, Philadelphia, PA, USA;
Marcelo Der Torossian Torres
Marcelo Der Torossian Torres
University of Pennsylvania
Peptide ChemistryAntimicrobial PeptidesPeptide Design
C
Cesar de la Fuente-Nunez
Machine Biology Group, Departments of Psychiatry and Microbiology, Institute for Biomedical Informatics, Institute for Translational Medicine and Therapeutics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, USA; Departments of Bioengineering and Chemical and Biomolecular Engineering, School of Engineering and Applied Science, University of Pennsylvania, Philadelphia, PA, USA; Department of Chemistry, School of Arts and Sciences, University of Pennsylvania, Philadelphia, PA, USA;
R
Ryan Marcus
Department of Computer and Information Science, University of Pennsylvania, Philadelphia, PA, USA
O
O. Bastani
Department of Computer and Information Science, University of Pennsylvania, Philadelphia, PA, USA
J
Jacob R. Gardner
Department of Computer and Information Science, University of Pennsylvania, Philadelphia, PA, USA