Multi-task Self-supervised Graph Neural Networks Enable Stronger Task Generalization

10/05/2022
by   Mingxuan Ju, et al.
0

Self-supervised learning (SSL) for graph neural networks (GNNs) has attracted increasing attention from the graph machine learning community in recent years, owing to its capability to learn performant node embeddings without costly label information. One weakness of conventional SSL frameworks for GNNs is that they learn through a single philosophy, such as mutual information maximization or generative reconstruction. When applied to various downstream tasks, these frameworks rarely perform equally well for every task, because one philosophy may not span the extensive knowledge required for all tasks. In light of this, we introduce ParetoGNN, a multi-task SSL framework for node representation learning over graphs. Specifically, ParetoGNN is self-supervised by manifold pretext tasks observing multiple philosophies. To reconcile different philosophies, we explore a multiple-gradient descent algorithm, such that ParetoGNN actively learns from every pretext task while minimizing potential conflicts. We conduct comprehensive experiments over four downstream tasks (i.e., node classification, node clustering, link prediction, and partition prediction), and our proposal achieves the best overall performance across tasks on 11 widely adopted benchmark datasets. Besides, we observe that learning from multiple philosophies enhances not only the task generalization but also the single task performance, demonstrating that ParetoGNN achieves better task generalization via the disjoint yet complementary knowledge learned from different philosophies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/01/2021

Self-supervised Auxiliary Learning for Graph Neural Networks via Meta-Learning

In recent years, graph neural networks (GNNs) have been widely adopted i...
research
07/19/2021

Adaptive Transfer Learning on Graph Neural Networks

Graph neural networks (GNNs) is widely used to learn a powerful represen...
research
08/03/2023

SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning

Textual graphs (TGs) are graphs whose nodes correspond to text (sentence...
research
12/12/2020

A Meta-Learning Approach for Graph Representation Learning in Multi-Task Settings

Graph Neural Networks (GNNs) are a framework for graph representation le...
research
01/10/2022

Graph Representation Learning for Multi-Task Settings: a Meta-Learning Approach

Graph Neural Networks (GNNs) have become the state-of-the-art method for...
research
06/23/2020

Gaining insight into SARS-CoV-2 infection and COVID-19 severity using self-supervised edge features and Graph Neural Networks

Graph Neural Networks (GNN) have been extensively used to extract meanin...
research
02/12/2023

USER: Unsupervised Structural Entropy-based Robust Graph Neural Network

Unsupervised/self-supervised graph neural networks (GNN) are vulnerable ...

Please sign up or login with your details

Forgot password? Click here to reset