🤖 AI Summary
To address coarse negative sample selection, insufficient robustness of visual representations, and the neglect of entity distinctiveness in multimodal entity linking (MEL), this paper proposes two core innovations: (1) Jaccard Distance–driven Conditional Contrastive Learning (JD-CCL), which enables attribute-aware hard negative mining by leveraging semantic similarity between entity descriptions; and (2) Context-Variant and Controllable Patch Transformation (CVaCPT), a context-aware mechanism that synthesizes multi-view images conditioned on textual context and dynamically modulates image patch representations to enhance fine-grained visual–semantic alignment. Evaluated on mainstream MEL benchmarks, our method achieves new state-of-the-art performance, improving linking accuracy by 3.2–5.8 percentage points. It demonstrates significant gains in fine-grained recognition and robustness under visually ambiguous conditions.
📝 Abstract
Previous research on multimodal entity linking (MEL) has primarily employed contrastive learning as the primary objective. However, using the rest of the batch as negative samples without careful consideration, these studies risk leveraging easy features and potentially overlook essential details that make entities unique. In this work, we propose JD-CCL (Jaccard Distance-based Conditional Contrastive Learning), a novel approach designed to enhance the ability to match multimodal entity linking models. JD-CCL leverages meta-information to select negative samples with similar attributes, making the linking task more challenging and robust. Additionally, to address the limitations caused by the variations within the visual modality among mentions and entities, we introduce a novel method, CVaCPT (Contextual Visual-aid Controllable Patch Transform). It enhances visual representations by incorporating multi-view synthetic images and contextual textual representations to scale and shift patch representations. Experimental results on benchmark MEL datasets demonstrate the strong effectiveness of our approach.