🤖 AI Summary
Existing CLIP-style multimodal contrastive pretraining frameworks fail to explicitly leverage the inherent linear structure of pointwise mutual information (PMI) in similarity modeling, leading to deviations from the theoretically optimal similarity metric. To address this, we propose KME-CLIP—the first framework to systematically incorporate PMI’s linearity into contrastive learning. Specifically, it models the optimal cross-modal similarity directly via kernel inner products in a reproducing kernel Hilbert space (RKHS), enabling arbitrary-precision approximation of the PMI-defined theoretical optimum without explicit PMI estimation. This design efficiently captures higher-order statistical dependencies. We provide rigorous theoretical guarantees on convergence and optimality. Empirically, KME-CLIP achieves significant improvements over standard CLIP on image–text retrieval and zero-shot classification, demonstrating both strong theoretical foundations and consistent performance gains.
📝 Abstract
In this study, we propose an enhancement to the similarity computation mechanism in multi-modal contrastive pretraining frameworks such as CLIP. Prior theoretical research has demonstrated that the optimal similarity metrics between paired modalities should correspond to the pointwise mutual information (PMI) between the two modalities. However, the current implementations of CLIP and its variants fail to fully utilize the underlying linear structure of PMI. We therefore propose KME-CLIP, which leverages this structure through the inner product in a reproducing kernel Hilbert space. We theoretically prove that our method can approximate PMI with arbitrary accuracy and empirically demonstrate that our approach overall outperforms the standard CLIP formulation across several retrieval and classification tasks.