🤖 AI Summary
Traditional self-supervised learning overly relies on data augmentation while neglecting semantic relationships among instances. To address this, we propose the first systematic incorporation of graph-structured modeling to explicitly capture inter-instance associations. Specifically, we construct a teacher–student dual-stream k-nearest neighbor (k-NN) graph and integrate graph neural networks (GNNs) to enable multi-hop message passing, thereby unifying local augmentations with global contextual information. Our approach introduces a k-NN-based dual-stream architecture coupled with a representation refinement mechanism, breaking away from the conventional paradigm that learns solely from intra-instance variations. Extensive experiments demonstrate consistent improvements in linear evaluation accuracy: +7.3% on CIFAR-10, +3.2% on ImageNet-100, and +1.0% on ImageNet-1K—outperforming state-of-the-art methods. These results validate the effectiveness and generalizability of explicitly modeling inter-instance relationships for self-supervised representation learning.
📝 Abstract
This paper introduces a novel approach that integrates graph theory into self-supervised representation learning. Traditional methods focus on intra-instance variations generated by applying augmentations. However, they often overlook important inter-instance relationships. While our method retains the intra-instance property, it further captures inter-instance relationships by constructing k-nearest neighbor (KNN) graphs for both teacher and student streams during pretraining. In these graphs, nodes represent samples along with their latent representations. Edges encode the similarity between instances. Following pretraining, a representation refinement phase is performed. In this phase, Graph Neural Networks (GNNs) propagate messages not only among immediate neighbors but also across multiple hops, thereby enabling broader contextual integration. Experimental results on CIFAR-10, ImageNet-100, and ImageNet-1K demonstrate accuracy improvements of 7.3%, 3.2%, and 1.0%, respectively, over state-of-the-art methods. These results highlight the effectiveness of the proposed graph based mechanism. The code is publicly available at https://github.com/alijavidani/SSL-GraphNNCLR.