Scalable Multi-Output Gaussian Processes with Stochastic Variational Inference

πŸ“… 2024-07-02
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Latent Variable Multi-Output Gaussian Processes (LV-MOGPs) suffer from linear computational complexity growth in the output dimension, severely limiting scalability to high-dimensional output settings. Method: We propose the first stochastic variational inference framework for LV-MOGPs supporting *dual batching*β€”simultaneous mini-batching over both inputs and outputs. Our approach enables scalable variational inference over output dimensions via kernelized output correlation modeling and a novel two-path mini-batch sampling strategy, reducing per-iteration complexity to output-dimension-independent. Contribution/Results: Theoretically grounded and empirically validated, our method accelerates training by over 10Γ— on 100-dimensional output tasks while preserving state-of-the-art predictive accuracy. Moreover, it supports zero-shot generalization to unseen outputs, substantially broadening the applicability of LV-MOGPs to large-scale multi-output regression.

Technology Category

Application Category

πŸ“ Abstract
The Multi-Output Gaussian Process is is a popular tool for modelling data from multiple sources. A typical choice to build a covariance function for a MOGP is the Linear Model of Coregionalization (LMC) which parametrically models the covariance between outputs. The Latent Variable MOGP (LV-MOGP) generalises this idea by modelling the covariance between outputs using a kernel applied to latent variables, one per output, leading to a flexible MOGP model that allows efficient generalization to new outputs with few data points. Computational complexity in LV-MOGP grows linearly with the number of outputs, which makes it unsuitable for problems with a large number of outputs. In this paper, we propose a stochastic variational inference approach for the LV-MOGP that allows mini-batches for both inputs and outputs, making computational complexity per training iteration independent of the number of outputs.
Problem

Research questions and friction points this paper is trying to address.

Scalable modeling of multi-output data with Gaussian Processes
Reducing computational complexity in Latent Variable MOGP
Enabling efficient generalization to new outputs with few data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Stochastic variational inference for LV-MOGP
Mini-batches for inputs and outputs
Scalable computation independent of output count
πŸ”Ž Similar Papers
Xiaoyu Jiang
Xiaoyu Jiang
Associate Professor (Research), Beihang University
Deep learningIndustrial IntelligenceAI security
S
Sokratia Georgaka
The University of Manchester
M
M. Rattray
The University of Manchester
M
Mauricio A. Alvarez
The University of Manchester