Variational Learning of Disentangled Representations

📅 2025-06-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In multi-condition representation learning, shared and condition-specific latent factors are often entangled, impairing model generalization across treatments, patients, or species. To address this, we propose DISCoVeR—a novel unsupervised framework featuring a dual-latent-variable architecture and an adversarial max-min optimization objective that enforces strict disentanglement without requiring manual priors. Theoretically, we prove the uniqueness of the equilibrium solution and its consistency with maximum-likelihood estimation. Methodologically, DISCoVeR integrates variational autoencoding, dual-path reconstruction, and information bottleneck regularization. Empirically, it achieves significant improvements in disentanglement metrics and cross-condition generalization on synthetic data, natural images, and single-cell RNA-seq datasets—demonstrating robust extraction of biologically stable signals.

Technology Category

Application Category

📝 Abstract
Disentangled representations enable models to separate factors of variation that are shared across experimental conditions from those that are condition-specific. This separation is essential in domains such as biomedical data analysis, where generalization to new treatments, patients, or species depends on isolating stable biological signals from context-dependent effects. While extensions of the variational autoencoder (VAE) framework have been proposed to address this problem, they frequently suffer from leakage between latent representations, limiting their ability to generalize to unseen conditions. Here, we introduce DISCoVeR, a new variational framework that explicitly separates condition-invariant and condition-specific factors. DISCoVeR integrates three key components: (i) a dual-latent architecture that models shared and specific factors separately; (ii) two parallel reconstructions that ensure both representations remain informative; and (iii) a novel max-min objective that encourages clean separation without relying on handcrafted priors, while making only minimal assumptions. Theoretically, we show that this objective maximizes data likelihood while promoting disentanglement, and that it admits a unique equilibrium. Empirically, we demonstrate that DISCoVeR achieves improved disentanglement on synthetic datasets, natural images, and single-cell RNA-seq data. Together, these results establish DISCoVeR as a principled approach for learning disentangled representations in multi-condition settings.
Problem

Research questions and friction points this paper is trying to address.

Separate shared and condition-specific factors in data
Prevent leakage between latent representations in VAEs
Improve generalization to unseen experimental conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-latent architecture separates shared and specific factors
Parallel reconstructions maintain informative representations
Max-min objective ensures clean separation without priors
🔎 Similar Papers
No similar papers found.
Yuli Slavutsky
Yuli Slavutsky
Postdoctoral Research Scientist, Columbia University
Machine LearningStatistics
O
Ozgur Beker
Department of Statistics, Columbia University, New York City, USA; Irving Institute for Cancer Dynamics, Columbia University, New York City, USA
David Blei
David Blei
Professor of Statistics and Computer Science, Columbia University
Machine LearningStatisticsProbabilistic topic modelsBayesian nonparametricsApproximate posterior inference
B
Bianca Dumitrascu
Department of Statistics, Columbia University, New York City, USA; Irving Institute for Cancer Dynamics, Columbia University, New York City, USA; Department of Computer Science, Columbia University, New York City, USA; Columbia Stem Cell Initiative, Columbia University, New York City, USA