Boosting Graph Neural Networks via Adaptive Knowledge Distillation

10/12/2022
by   Zhichun Guo, et al.
0

Graph neural networks (GNNs) have shown remarkable performance on diverse graph mining tasks. Although different GNNs can be unified as the same message passing framework, they learn complementary knowledge from the same graph. Knowledge distillation (KD) is developed to combine the diverse knowledge from multiple models. It transfers knowledge from high-capacity teachers to a lightweight student. However, to avoid oversmoothing, GNNs are often shallow, which deviates from the setting of KD. In this context, we revisit KD by separating its benefits from model compression and emphasizing its power of transferring knowledge. To this end, we need to tackle two challenges: how to transfer knowledge from compact teachers to a student with the same capacity; and, how to exploit student GNN's own strength to learn knowledge. In this paper, we propose a novel adaptive KD framework, called BGNN, which sequentially transfers knowledge from multiple GNNs into a student GNN. We also introduce an adaptive temperature module and a weight boosting module. These modules guide the student to the appropriate knowledge for effective learning. Extensive experiments have demonstrated the effectiveness of BGNN. In particular, we achieve up to 3.05 7.67

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/24/2022

Geometric Knowledge Distillation: Topology Compression for Graph Neural Networks

We study a new paradigm of knowledge transfer that aims at encoding grap...
research
01/03/2023

RELIANT: Fair Knowledge Distillation for Graph Neural Networks

Graph Neural Networks (GNNs) have shown satisfying performance on variou...
research
08/18/2023

Bridged-GNN: Knowledge Bridge Learning for Effective Knowledge Transfer

The data-hungry problem, characterized by insufficiency and low-quality ...
research
08/04/2023

VQGraph: Graph Vector-Quantization for Bridging GNNs and MLPs

Graph Neural Networks (GNNs) conduct message passing which aggregates lo...
research
06/10/2021

Graph Symbiosis Learning

We introduce a framework for learning from multiple generated graph view...
research
08/12/2021

Distilling Holistic Knowledge with Graph Neural Networks

Knowledge Distillation (KD) aims at transferring knowledge from a larger...
research
02/01/2023

Knowledge Distillation on Graphs: A Survey

Graph Neural Networks (GNNs) have attracted tremendous attention by demo...

Please sign up or login with your details

Forgot password? Click here to reset