Probabilistic Variational Contrastive Learning

πŸ“… 2025-06-11
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing contrastive learning methods (e.g., SimCLR, SupCon) produce deterministic embeddings without uncertainty quantification. To address this, we propose VCLβ€”the first decoder-free variational contrastive learning framework. VCL interprets the InfoNCE loss as a surrogate reconstruction term, imposes a uniform prior on the unit hypersphere, and models the posterior via a projected normal distribution to enable sampling-based probabilistic embeddings. We further introduce a normalized KL regularization term to effectively mitigate dimensional collapse. This work establishes the first rigorous variational inference foundation for contrastive learning. Experiments demonstrate that VCL matches or exceeds the classification accuracy of SimCLR and SupCon across multiple benchmarks, significantly increases mutual information between embeddings and class labels, and provides interpretable, sample-based uncertainty estimates for learned representations.

Technology Category

Application Category

πŸ“ Abstract
Deterministic embeddings learned by contrastive learning (CL) methods such as SimCLR and SupCon achieve state-of-the-art performance but lack a principled mechanism for uncertainty quantification. We propose Variational Contrastive Learning (VCL), a decoder-free framework that maximizes the evidence lower bound (ELBO) by interpreting the InfoNCE loss as a surrogate reconstruction term and adding a KL divergence regularizer to a uniform prior on the unit hypersphere. We model the approximate posterior $q_ heta(z|x)$ as a projected normal distribution, enabling the sampling of probabilistic embeddings. Our two instantiations--VSimCLR and VSupCon--replace deterministic embeddings with samples from $q_ heta(z|x)$ and incorporate a normalized KL term into the loss. Experiments on multiple benchmarks demonstrate that VCL mitigates dimensional collapse, enhances mutual information with class labels, and matches or outperforms deterministic baselines in classification accuracy, all the while providing meaningful uncertainty estimates through the posterior model. VCL thus equips contrastive learning with a probabilistic foundation, serving as a new basis for contrastive approaches.
Problem

Research questions and friction points this paper is trying to address.

Lack of uncertainty quantification in contrastive learning embeddings
Need for probabilistic embeddings in contrastive learning frameworks
Dimensional collapse and mutual information enhancement in embeddings
Innovation

Methods, ideas, or system contributions that make the work stand out.

VCL maximizes ELBO with InfoNCE and KL divergence
Models posterior as projected normal distribution
Replaces deterministic embeddings with probabilistic samples
πŸ”Ž Similar Papers
No similar papers found.