Invariant Representations via Wasserstein Correlation Maximization

📅 2025-05-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the challenge of jointly preserving geometric fidelity and enhancing robustness in unsupervised representation learning. We propose a novel framework based on Wasserstein correlation: by maximizing the Wasserstein correlation between the input data distribution and the encoded distribution, our autoencoder preserves the topological and geometric structure of the original data during dimensionality reduction. To our knowledge, this is the first work to leverage Wasserstein correlation for learning representations that are approximately invariant to specified augmentations (or augmentation sets). We introduce a Markov–Wasserstein kernel to construct an augmentation encoder, enabling efficient acquisition of invariance without fine-tuning—i.e., with the pretrained backbone frozen. We provide theoretical guarantees establishing the optimality and stability of the proposed metric. Empirical results demonstrate that even simple feedforward networks trained under this framework yield highly invariant and geometrically faithful representations.

Technology Category

Application Category

📝 Abstract
This work investigates the use of Wasserstein correlation -- a normalized measure of statistical dependence based on the Wasserstein distance between a joint distribution and the product of its marginals -- for unsupervised representation learning. Unlike, for example, contrastive methods, which naturally cluster classes in the latent space, we find that an (auto)encoder trained to maximize Wasserstein correlation between the input and encoded distributions instead acts as a compressor, reducing dimensionality while approximately preserving the topological and geometric properties of the input distribution. More strikingly, we show that Wasserstein correlation maximization can be used to arrive at an (auto)encoder -- either trained from scratch, or else one that extends a frozen, pretrained model -- that is approximately invariant to a chosen augmentation, or collection of augmentations, and that still approximately preserves the structural properties of the non-augmented input distribution. To do this, we first define the notion of an augmented encoder using the machinery of Markov-Wasserstein kernels. When the maximization objective is then applied to the augmented encoder, as opposed to the underlying, deterministic encoder, the resulting model exhibits the desired invariance properties. Finally, besides our experimental results, which show that even simple feedforward networks can be imbued with invariants or can, alternatively, be used to impart invariants to pretrained models under this training process, we additionally establish various theoretical results for optimal transport-based dependence measures. Code is available at https://github.com/keenan-eikenberry/wasserstein_correlation_maximization .
Problem

Research questions and friction points this paper is trying to address.

Maximizing Wasserstein correlation for unsupervised representation learning
Creating invariant encoders preserving input distribution properties
Extending pretrained models with augmentation-invariant features
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Wasserstein correlation for unsupervised representation learning
Maximizes correlation to preserve input distribution properties
Defines augmented encoder with Markov-Wasserstein kernels
🔎 Similar Papers
No similar papers found.
K
Keenan Eikenberry
Department of Mathematics, Dartmouth College
Lizuo Liu
Lizuo Liu
Postdoctoral Research Associate
Applied Math
Y
Yoonsang Lee
Department of Mathematics, Dartmouth College