🤖 AI Summary
This paper addresses the challenge of uncertainty quantification for regression models in metric spaces, where outputs are random objects such as probability distributions or graph Laplacian matrices. Methodologically, it introduces: (1) a novel definition of homoscedasticity in metric spaces, enabling finite-sample conformal predictors with rigorous coverage guarantees; (2) a calibration-free, local k-nearest-neighbor adaptive algorithm for heteroscedastic settings, balancing computational efficiency with preservation of underlying geometric structure; and (3) a synthesis of nonparametric metric-space regression and conformal inference, ensuring estimator consistency under minimal assumptions. The proposed framework is the first conformal prediction method applicable to arbitrary black-box regressors in metric spaces—scalable, theoretically grounded, and assumption-light. Experiments demonstrate its effectiveness and scalability on real-world applications, including personalized medicine.
📝 Abstract
This paper introduces a framework for uncertainty quantification in regression models defined in metric spaces. Leveraging a newly defined notion of homoscedasticity, we develop a conformal prediction algorithm that offers finite-sample coverage guarantees and fast convergence rates of the oracle estimator. In heteroscedastic settings, we forgo these non-asymptotic guarantees to gain statistical efficiency, proposing a local $k$--nearest--neighbor method without conformal calibration that is adaptive to the geometry of each particular nonlinear space. Both procedures work with any regression algorithm and are scalable to large data sets, allowing practitioners to plug in their preferred models and incorporate domain expertise. We prove consistency for the proposed estimators under minimal conditions. Finally, we demonstrate the practical utility of our approach in personalized--medicine applications involving random response objects such as probability distributions and graph Laplacians.