TY - JOUR
T1 - Beyond Augmentation
T2 - Leveraging Inter-Instance Relation in Self-Supervised Representation Learning
AU - Javidani, Ali
AU - Araabi, Babak Nadjar
AU - Sadeghi, Mohammad Amin
N1 - Publisher Copyright:
© 1994-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - This letter introduces a novel approach that integrates graph theory into self-supervised representation learning. Traditional methods focus on intra-instance variations generated by applying augmentations. However, they often overlook important inter-instance relationships. While our method retains the intra-instance property, it further captures inter-instance relationships by constructing k -nearest neighbor (KNN) graphs for both teacher and student streams during pretraining. In these graphs, nodes represent samples along with their latent representations. Edges encode the similarity between instances. Following pretraining, a representation refinement phase is performed. In this phase, Graph Neural Networks (GNNs) propagate messages not only among immediate neighbors but also across multiple hops, thereby enabling broader contextual integration. Experimental results on CIFAR-10, ImageNet-100, and ImageNet-1K demonstrate accuracy improvements of 7.3%, 3.2%, and 1.0%, respectively, over state-of-the-art methods. These results highlight the effectiveness of the proposed graph-based mechanism.
AB - This letter introduces a novel approach that integrates graph theory into self-supervised representation learning. Traditional methods focus on intra-instance variations generated by applying augmentations. However, they often overlook important inter-instance relationships. While our method retains the intra-instance property, it further captures inter-instance relationships by constructing k -nearest neighbor (KNN) graphs for both teacher and student streams during pretraining. In these graphs, nodes represent samples along with their latent representations. Edges encode the similarity between instances. Following pretraining, a representation refinement phase is performed. In this phase, Graph Neural Networks (GNNs) propagate messages not only among immediate neighbors but also across multiple hops, thereby enabling broader contextual integration. Experimental results on CIFAR-10, ImageNet-100, and ImageNet-1K demonstrate accuracy improvements of 7.3%, 3.2%, and 1.0%, respectively, over state-of-the-art methods. These results highlight the effectiveness of the proposed graph-based mechanism.
KW - Self-supervised learning
KW - graph neural networks
KW - k-nearest neighbor graph
KW - representation learning
UR - https://www.scopus.com/pages/publications/105016469979
U2 - 10.1109/LSP.2025.3610549
DO - 10.1109/LSP.2025.3610549
M3 - Article
AN - SCOPUS:105016469979
SN - 1070-9908
VL - 32
SP - 3730
EP - 3734
JO - IEEE Signal Processing Letters
JF - IEEE Signal Processing Letters
ER -