Learning structure perception MLPs on graphs: a layer-wise graph knowledge distillation framework

被引:0
|
作者
Du, Hangyuan [1 ]
Yu, Rong [1 ]
Bai, Liang [2 ,3 ]
Bai, Lu [4 ,5 ]
Wang, Wenjian [2 ]
机构
[1] Shanxi Univ, Sch Comp & Informat Technol, Taiyuan 030006, Shanxi, Peoples R China
[2] Shanxi Univ, Key Lab Computat Intelligence Chinese Informat Pro, Minist Educ, Taiyuan 030006, Shanxi, Peoples R China
[3] Shanxi Univ, Inst Intelligent Informat Proc, Taiyuan 030006, Shanxi, Peoples R China
[4] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
[5] Cent Univ Finance & Econ, Beijing 100875, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph knowledge distillation; Supervision signal; Layer-wise mapping; Structure perception MLPs;
D O I
10.1007/s13042-024-02150-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) are expressive in dealing with graph data. Because of the large storage requirements and the high computational complexity, it is difficult to deploy these cumbersome models in resource-constrained environments. As a representative model compression strategy, knowledge distillation (KD) is introduced into graph analysis research to address this problem. However, there are some crucial challenges in existing graph knowledge distillation algorithms, such as knowledge transfer effectiveness and student model designation. To address these problems, a new graph distillation model is proposed in this paper. Specifically, a layer-wise mapping strategy is designed to distill knowledge for training the student model, in which staged knowledge learned by intermediate layers of teacher GNNs is captured to form supervision signals. And, an adaptive weight mechanism is developed to evaluate the importance of the distilled knowledge. On this basis, a structure perception MLPs is constructed as the student model, which can capture prior information of the input graph from the perspectives of node feature and topology structure. In this way, the proposed model shares the prediction advantage of GNNs and the latency advantage of MLPs. Node classification experiments on five benchmark datasets demonstrate the validity and superiority of our model over baseline algorithms.
引用
收藏
页码:4357 / 4372
页数:16
相关论文
共 50 条
  • [1] Layer-wise Knowledge Distillation for Cross-Device Federated Learning
    Le, Huy Q.
    Nguyen, Loc X.
    Park, Seong-Bae
    Hong, Choong Seon
    [J]. 2023 INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN, 2023, : 526 - 529
  • [2] Decoupled graph knowledge distillation: A general logits-based method for learning MLPs on graphs
    Tian, Yingjie
    Xu, Shaokai
    Li, Muyang
    [J]. NEURAL NETWORKS, 2024, 179
  • [3] Craft Distillation: Layer-wise Convolutional Neural Network Distillation
    Blakeney, Cody
    Li, Xiaomin
    Yan, Yan
    Zong, Ziliang
    [J]. 2020 7TH IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND CLOUD COMPUTING (CSCLOUD 2020)/2020 6TH IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING AND SCALABLE CLOUD (EDGECOM 2020), 2020, : 252 - 257
  • [4] An analytic layer-wise deep learning framework with applications to robotics
    Huu-Thiet Nguyen
    Chien Chern Cheah
    Kar-Ann Toh
    [J]. AUTOMATICA, 2022, 135
  • [5] A Layer-Wise Theoretical Framework for Deep Learning of Convolutional Neural Networks
    Huu-Thiet Nguyen
    Li, Sitan
    Cheah, Chien Chern
    [J]. IEEE ACCESS, 2022, 10 : 14270 - 14287
  • [6] DISTILHUBERT: SPEECH REPRESENTATION LEARNING BY LAYER-WISE DISTILLATION OF HIDDEN-UNIT BERT
    Chang, Heng-Jui
    Yang, Shu-wen
    Lee, Hung-yi
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7087 - 7091
  • [7] LAD: Layer-Wise Adaptive Distillation for BERT Model Compression
    Lin, Ying-Jia
    Chen, Kuan-Yu
    Kao, Hung-Yu
    [J]. SENSORS, 2023, 23 (03)
  • [8] Layer-Wise Learning Framework for Efficient DNN Deployment in Biomedical Wearable Systems
    [J]. 2023 IEEE 19TH INTERNATIONAL CONFERENCE ON BODY SENSOR NETWORKS, BSN, 2023,
  • [9] Deep Learning Layer-Wise Learning of Feature Hierarchies
    Schulz, Hannes
    Behnke, Sven
    [J]. KUNSTLICHE INTELLIGENZ, 2012, 26 (04): : 357 - 363
  • [10] FedLF: Layer-Wise Fair Federated Learning
    Pan, Zibin
    Li, Chi
    Yu, Fangchen
    Wang, Shuyi
    Wang, Haijin
    Tang, Xiaoying
    Zhao, Junhua
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14527 - 14535