Dual-channel graph contrastive learning for self-supervised graph-level representation learning

Research output: Contribution to journalArticlepeer-review

25 Scopus citations

Abstract

Self-supervised graph-level representation learning aims to learn discriminative representations for subgraphs or entire graphs without human-curated labels. Recently, graph contrastive learning (GCL) methods have revolutionized this field and achieved state-of-the-art results in various downstream tasks. Nonetheless, current GCL models are mostly based on simple node-level information aggregation operations and fail to reveal various substructures from input graphs. Moreover, to perform graph-graph contrastive training, they often involve well-designed graph augmentation, which is expensive and requires extensive expert efforts. Here, we propose a novel GCL framework, namely DualGCL, for self-supervised graph-level representation learning. For fine-grained local information incorporation, we first present an adaptive hierarchical aggregation process with a differentiable Transformer-based aggregator. Then, to efficiently learn graph-level discriminative representations, we introduce a dual-channel contrastive learning process in a multi-granularity and augmentation-free contrasting mode. When tested empirically on six popular graph classification benchmarks, our DualGCL achieves better or comparable performance than various strong baselines.

Original languageEnglish
Article number109448
JournalPattern Recognition
Volume139
DOIs
StatePublished - Jul 2023

Keywords

  • Contrastive learning
  • Graph classification
  • Graph neural networks
  • Graph representation learning

Fingerprint

Dive into the research topics of 'Dual-channel graph contrastive learning for self-supervised graph-level representation learning'. Together they form a unique fingerprint.

Cite this