🤖 AI Summary
Existing neural optimal transport (OT) methods are limited to modeling single distribution pairs, exhibiting poor generalization across multiple empirical distributions. Method: We propose the first generalized neural OT framework capable of learning OT mappings for arbitrary unseen empirical distributions. Our approach introduces a Transformer-based encoder to embed variable-length discrete distributions and designs a neural hypernetwork that dynamically generates distribution-pair-specific OT mappings—thereby departing from conventional pairwise modeling paradigms. Contribution/Results: By end-to-end optimizing the Wasserstein distance, our method achieves high-accuracy, low-latency OT mapping generation across diverse synthetic and real-world benchmarks, significantly outperforming state-of-the-art baselines. The implementation is publicly available.
📝 Abstract
Distributional data have become increasingly prominent in modern signal processing, highlighting the necessity of computing optimal transport (OT) maps across multiple probability distributions. Nevertheless, recent studies on neural OT methods predominantly focused on the efficient computation of a single map between two distributions. To address this challenge, we introduce a novel approach to learning transport maps for new empirical distributions. Specifically, we employ the transformer architecture to produce embeddings from distributional data of varying length; these embeddings are then fed into a hypernetwork to generate neural OT maps. Various numerical experiments were conducted to validate the embeddings and the generated OT maps. The model implementation and the code are provided on https://github.com/jiangmingchen/HOTET.