Homophilic-aware graph contrastive learning

Li Zhang, Hua Mao, Wai Lok Woo, Jie Chen*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Graph contrastive learning (GCL) has become increasingly popular in unsupervised graph representation learning (UGRL). Currently, most existing GCL approaches implicitly rely on the homophily assumption. These approaches encounter two challenges when adapting to heterophily issues. First, they often suffer from losing intrinsic structural information owing to the separate encoding processes that are applied to homophilic and heterophilic views. Second, they overlook the importance of different subgraphs when disentangling a heterophilic graph into several relational graphs. In this paper, we propose a homophilic-aware GCL (HAGCL) method that learns informative node representations for both homophilic and heterophilic graphs in a unified framework. Specifically, we first introduce a homophilic-aware augmentation scheme for generating more effective views with a higher homophily ratio. This scheme potentially enables the learned node representations to be more informative and robust. We then introduce a discriminative relation and frequency adaptation (DRFA) module to address the disentangling problem. This DRFA module can adaptively capture diverse relations and frequency signals while discriminatively fusing multiple relational graphs. Extensive experimental results that were obtained on homophilic and heterophilic datasets demonstrate the effectiveness of the proposed HAGCL method.
Original languageEnglish
Article number113328
Pages (from-to)1-13
Number of pages13
JournalPattern Recognition
Volume177
Early online date16 Feb 2026
DOIs
Publication statusE-pub ahead of print - 16 Feb 2026

Keywords

  • Contrastive learning
  • Graph neural networks
  • Heterophilic graph
  • Node classification

Fingerprint

Dive into the research topics of 'Homophilic-aware graph contrastive learning'. Together they form a unique fingerprint.

Cite this