🤖 AI Summary
This work proposes a novel paradigm for flexible conditional density estimation by modeling the cumulative distribution function (CDF) as the primary target instead of directly estimating the probability density function (PDF). Direct PDF modeling often suffers from ill-posedness and performance degradation under multimodal, asymmetric, or topologically complex distributions. By learning a smooth CDF and differentiating it to obtain the PDF, the approach effectively mitigates noise amplification. The authors introduce the Smooth Min-Max (SMM) neural network architecture, which combines differentiable smoothing operations with autoregressive decomposition to guarantee valid PDFs and enable multivariate modeling, facilitating approximate likelihood-based training. Experimental results demonstrate that the method consistently outperforms state-of-the-art approaches in both univariate and multivariate settings, achieving significantly improved estimation accuracy and numerical stability.
📝 Abstract
Conditional density estimation (CDE) is a fundamental task in machine learning that aims to model the full conditional law $\mathbb{P}(\mathbf{y} \mid \mathbf{x})$, beyond mere point prediction (e.g., mean, mode). A core challenge is free-form density estimation, capturing distributions that exhibit multimodality, asymmetry, or topological complexity without restrictive assumptions. However, prevailing methods typically estimate the probability density function (PDF) directly, which is mathematically ill-posed: differentiating the empirical distribution amplifies random fluctuations inherent in finite datasets, necessitating strong inductive biases that limit expressivity and fail when violated. We propose a CDF-first framework that circumvents this issue by estimating the cumulative distribution function (CDF), a stable and well-posed target, and then recovering the PDF via differentiation of the learned smooth CDF. Parameterizing the CDF with a Smooth Min-Max (SMM) network, our framework guarantees valid PDFs by construction, enables tractable approximate likelihood training, and preserves complex distributional shapes. For multivariate outputs, we use an autoregressive decomposition with SMM factors. Experiments demonstrate our approach outperforms state-of-the-art density estimators on a range of univariate and multivariate tasks.