🤖 AI Summary
This study addresses the challenge of cross-modal alignment and understanding among heterogeneous geospatial data sources—such as aerial imagery, street views, elevation models, textual descriptions, and geographic coordinates—by proposing a unified multimodal contrastive learning framework. Departing from conventional centralized fusion strategies, the framework employs an all-to-all contrastive alignment mechanism and incorporates a multi-scale latitude–longitude encoder to accurately capture geographic structure, thereby mapping all five modalities into a shared embedding space. Experimental results demonstrate that the proposed approach significantly outperforms both single-modality models and coordinate-only baselines across multiple downstream geospatial tasks, confirming its effectiveness and superiority in cross-modal retrieval, reasoning, and representation learning.
📝 Abstract
The growing availability of co-located geospatial data spanning aerial imagery, street-level views, elevation models, text, and geographic coordinates offers a unique opportunity for multimodal representation learning. We introduce UNIGEOCLIP, a massively multimodal contrastive framework to jointly align five complementary geospatial modalities in a single unified embedding space. Unlike prior approaches that fuse modalities or rely on a central pivot representation, our method performs all-to-all contrastive alignment, enabling seamless comparison, retrieval, and reasoning across arbitrary combinations of modalities. We further propose a scaled latitude-longitude encoder that improves spatial representation by capturing multi-scale geographic structure. Extensive experiments across downstream geospatial tasks demonstrate that UNIGEOCLIP consistently outperforms single-modality contrastive models and coordinate-only baselines, highlighting the benefits of holistic multimodal geospatial alignment. A reference implementation is available at https://gastruc.github.io/unigeoclip.