๐ค AI Summary
To address the computational intractability of exact Gaussian process (GP) inference in large-scale GP variational autoencoders (GPVAEs), this paper proposes a neighborhood-driven local approximation method: for each data point, GP inference is performed only over its *k*-nearest neighbors in the latent spaceโbypassing global kernel assumptions and eliminating the need for large sets of inducing points. This is the first work to incorporate neighborhood locality into the GPVAE variational inference framework, enabling more flexible kernel design while preserving essential latent variable dependencies. Experiments demonstrate substantial improvements over existing GPVAE baselines across representation learning, missing data imputation, and conditional generation tasks: prediction accuracy increases significantly, training speed accelerates by 3โ5ร, and memory consumption decreases by approximately 40%.
๐ Abstract
Gaussian Process (GP) Variational Autoencoders (VAEs) extend standard VAEs by replacing the fully factorised Gaussian prior with a GP prior, thereby capturing richer correlations among latent variables. However, performing exact GP inference in large-scale GPVAEs is computationally prohibitive, often forcing existing approaches to rely on restrictive kernel assumptions or large sets of inducing points. In this work, we propose a neighbour-driven approximation strategy that exploits local adjacencies in the latent space to achieve scalable GPVAE inference. By confining computations to the nearest neighbours of each data point, our method preserves essential latent dependencies, allowing more flexible kernel choices and mitigating the need for numerous inducing points. Through extensive experiments on tasks including representation learning, data imputation, and conditional generation, we demonstrate that our approach outperforms other GPVAE variants in both predictive performance and computational efficiency.