HiGP: A high-performance Python package for Gaussian Process

📅 2025-03-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Standard Gaussian processes (GPs) suffer from prohibitive computational complexity and poor scalability on large-scale datasets (thousands to millions of instances). This paper proposes an efficient GP regression and classification framework addressing these limitations. Methodologically, it introduces (1) a novel preconditioning strategy that integrates adaptive Nyström decomposition with customized kernel matrix–vector and matrix–matrix multiplication routines, substantially accelerating convergence and improving numerical stability of iterative solvers such as conjugate gradient; and (2) a PyTorch-based implementation supporting exact gradient computation, multiple kernel acceleration techniques, and end-to-end uncertainty quantification. Empirical results demonstrate that the proposed method achieves 10×–100× speedups in both training and inference over standard GP implementations while preserving modeling accuracy. The framework thus provides a scalable, high-fidelity, and general-purpose solution for large-scale GP applications.

Technology Category

Application Category

📝 Abstract
Gaussian Processes (GPs) are flexible, nonparametric Bayesian models widely used for regression and classification tasks due to their ability to capture complex data patterns and provide uncertainty quantification (UQ). Traditional GP implementations often face challenges in scalability and computational efficiency, especially with large datasets. To address these challenges, HiGP, a high-performance Python package, is designed for efficient Gaussian Process regression (GPR) and classification (GPC) across datasets of varying sizes. HiGP combines multiple new iterative methods to enhance the performance and efficiency of GP computations. It implements various effective matrix-vector (MatVec) and matrix-matrix (MatMul) multiplication strategies specifically tailored for kernel matrices. To improve the convergence of iterative methods, HiGP also integrates the recently developed Adaptive Factorized Nystrom (AFN) preconditioner and employs precise formulas for computing the gradients. With a user-friendly Python interface, HiGP seamlessly integrates with PyTorch and other Python packages, allowing easy incorporation into existing machine learning and data analysis workflows.
Problem

Research questions and friction points this paper is trying to address.

Addresses scalability and computational efficiency in Gaussian Process models.
Enhances performance of GP regression and classification for large datasets.
Integrates advanced methods for efficient matrix operations and convergence.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Efficient Gaussian Process regression and classification
Advanced matrix-vector and matrix-multiplication strategies
Integration of Adaptive Factorized Nystrom preconditioner
🔎 Similar Papers
No similar papers found.
H
Hua Huang
School of Computational Science and Engineering, Georgia Institute of Technology, Atlanta, GA
T
Tianshi Xu
Department of Mathematics, Emory University, Atlanta, GA
Yuanzhe Xi
Yuanzhe Xi
Associate Professor, Emory University
Numerical linear algebraScientific Machine Learning
Edmond Chow
Edmond Chow
Georgia Institute of Technology
scientific computinghigh-performance computingnumerical methods