Generative Modeling of Discrete Data Using Geometric Latent Subspaces

📅 2026-01-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges of modeling complex dependencies and mitigating redundancy in high-dimensional parameter spaces for discrete data generation. It introduces, for the first time, a Riemannian geometric structure with isometric properties into the exponential parameter space of product manifolds over categorical distributions, thereby constructing a low-dimensional latent subspace. By leveraging the Riemannian metric, geodesics within this subspace become straight lines, enabling consistent and efficient flow-matching training. The proposed approach substantially reduces the dimensionality of latent variables while preserving strong representational capacity for discrete data distributions. Experimental results demonstrate that the model achieves accurate and efficient discrete data generation using a significantly lower-dimensional latent space, effectively balancing computational efficiency with modeling performance.

Technology Category

Application Category

📝 Abstract
We introduce the use of latent subspaces in the exponential parameter space of product manifolds of categorial distributions, as a tool for learning generative models of discrete data. The low-dimensional latent space encodes statistical dependencies and removes redundant degrees of freedom among the categorial variables. We equip the parameter domain with a Riemannian geometry such that the spaces and distances are related by isometries which enables consistent flow matching. In particular, geodesics become straight lines which makes model training by flow matching effective. Empirical results demonstrate that reduced latent dimensions suffice to represent data for generative modeling.
Problem

Research questions and friction points this paper is trying to address.

generative modeling
discrete data
latent subspaces
categorical distributions
statistical dependencies
Innovation

Methods, ideas, or system contributions that make the work stand out.

latent subspaces
discrete generative modeling
Riemannian geometry
flow matching
categorical distributions
🔎 Similar Papers
No similar papers found.
Daniel Gonzalez-Alvarado
Daniel Gonzalez-Alvarado
Heidelberg University
Uncertainty QuantificationGenerative ModelingNeural ODEs
J
Jonas Cassel
Institute for Mathematics, Heidelberg University; Research Station Geometry and Dynamics, Heidelberg University
S
Stefania Petra
Institute for Mathematics, Heidelberg University; Mannheim Institute for Intelligent Systems in Medicine, Heidelberg University; IWR, Heidelberg University
C
Christoph Schnorr
Institute for Mathematics, Heidelberg University; IWR, Heidelberg University; Research Station Geometry and Dynamics, Heidelberg University