🤖 AI Summary
Nonparametric modeling of complex multivariate data remains challenging due to the difficulty of simultaneously capturing heterogeneous marginal distributions, high-order dependence structures, and conditional independence relationships.
Method: We propose a novel framework that replaces the conventional Gaussian copula with a customized multivariate nonlinear transformation. Flexible dependence structures are modeled via penalized splines, and a Lasso-type sparse regularization is introduced—enabling, for the first time, direct learning of pairwise conditional independencies and subsequent graphical structure inference.
Contribution/Results: In simulations, our method accurately recovers parametric vine copulas and correctly identifies conditional independence structures. On astrophysical benchmark data, it significantly outperforms existing nonparametric vine copula approaches, achieving superior accuracy in both distributional fitting and dependence structure recovery. The framework uniquely balances high-fidelity distributional approximation with statistically interpretable graphical inference.
📝 Abstract
Graphical Transformation Models (GTMs) are introduced as a novel approach to effectively model multivariate data with intricate marginals and complex dependency structures non-parametrically, while maintaining interpretability through the identification of varying conditional independencies. GTMs extend multivariate transformation models by replacing the Gaussian copula with a custom-designed multivariate transformation, offering two major advantages. Firstly, GTMs can capture more complex interdependencies using penalized splines, which also provide an efficient regularization scheme. Secondly, we demonstrate how to approximately regularize GTMs using a lasso penalty towards pairwise conditional independencies, akin to Gaussian graphical models. The model's robustness and effectiveness are validated through simulations, showcasing its ability to accurately learn parametric vine copulas and identify conditional independencies. Additionally, the model is applied to a benchmark astrophysics dataset, where the GTM demonstrates favorable performance compared to non-parametric vine copulas in learning complex multivariate distributions.