Learning structure perception MLPs on graphs: a layer-wise graph knowledge distillation framework

被引:0
|
作者
Du, Hangyuan [1 ]
Yu, Rong [1 ]
Bai, Liang [2 ,3 ]
Bai, Lu [4 ,5 ]
Wang, Wenjian [2 ]
机构
[1] Shanxi Univ, Sch Comp & Informat Technol, Taiyuan 030006, Shanxi, Peoples R China
[2] Shanxi Univ, Key Lab Computat Intelligence Chinese Informat Pro, Minist Educ, Taiyuan 030006, Shanxi, Peoples R China
[3] Shanxi Univ, Inst Intelligent Informat Proc, Taiyuan 030006, Shanxi, Peoples R China
[4] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
[5] Cent Univ Finance & Econ, Beijing 100875, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph knowledge distillation; Supervision signal; Layer-wise mapping; Structure perception MLPs;
D O I
10.1007/s13042-024-02150-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) are expressive in dealing with graph data. Because of the large storage requirements and the high computational complexity, it is difficult to deploy these cumbersome models in resource-constrained environments. As a representative model compression strategy, knowledge distillation (KD) is introduced into graph analysis research to address this problem. However, there are some crucial challenges in existing graph knowledge distillation algorithms, such as knowledge transfer effectiveness and student model designation. To address these problems, a new graph distillation model is proposed in this paper. Specifically, a layer-wise mapping strategy is designed to distill knowledge for training the student model, in which staged knowledge learned by intermediate layers of teacher GNNs is captured to form supervision signals. And, an adaptive weight mechanism is developed to evaluate the importance of the distilled knowledge. On this basis, a structure perception MLPs is constructed as the student model, which can capture prior information of the input graph from the perspectives of node feature and topology structure. In this way, the proposed model shares the prediction advantage of GNNs and the latency advantage of MLPs. Node classification experiments on five benchmark datasets demonstrate the validity and superiority of our model over baseline algorithms.
引用
收藏
页码:4357 / 4372
页数:16
相关论文
共 50 条
  • [31] Layer-Wise Personalized Federated Learning for Mobile Traffic Prediction
    Lee, Seungyeol
    Sung, Jihoon
    Shin, Myung-Ki
    [J]. IEEE ACCESS, 2024, 12 : 53126 - 53140
  • [32] Collaborative Layer-Wise Discriminative Learning in Deep Neural Networks
    Jin, Xiaojie
    Chen, Yunpeng
    Dong, Jian
    Feng, Jiashi
    Yan, Shuicheng
    [J]. COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 : 733 - 749
  • [33] Enriching Variety of Layer-wise Learning Information by Gradient Combination
    Wang, Chien-Yao
    Liao, Hong-Yuan Mark
    Chen, Ping-Yang
    Hsieh, Jun-Wei
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 2477 - 2484
  • [34] A Layer-wise Score Level Ensemble Framework for Acoustic Scene Classification
    Singh, Arshdeep
    Thakur, Anshul
    Rajan, Padmanabhan
    Bhavsar, Arnav
    [J]. 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 837 - 841
  • [35] Semantic Stage-Wise Learning for Knowledge Distillation
    Liu, Dongqin
    Li, Wentao
    Zhou, Wei
    Li, Zhaoxing
    Dai, Jiao
    Han, Jizhong
    Li, Ruixuan
    Hu, Songlin
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 816 - 821
  • [36] Efficient Federated Learning Using Layer-Wise Regulation and Momentum Aggregation*
    Zhang, Fan
    Fang, Zekuan
    Li, Yiming
    Chen, Mingsong
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (18)
  • [37] Learning Feature Hierarchies: A Layer-Wise Tag-Embedded Approach
    Yuan, Zhaoquan
    Xu, Changsheng
    Sang, Jitao
    Yan, Shuicheng
    Hossain, M. Shamim
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2015, 17 (06) : 816 - 827
  • [38] Guided Layer-Wise Learning for Deep Models Using Side Information
    Sulimov, Pavel
    Sukmanova, Elena
    Chereshnev, Roman
    Kertesz-Farkas, Attila
    [J]. ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS (AIST 2019), 2020, 1086 : 50 - 61
  • [39] YOGA: Adaptive Layer-Wise Model Aggregation for Decentralized Federated Learning
    Liu, Jun
    Liu, Jianchun
    Xu, Hongli
    Liao, Yunming
    Wang, Zhiyuan
    Ma, Qianpiao
    [J]. IEEE-ACM TRANSACTIONS ON NETWORKING, 2024, 32 (02) : 1768 - 1780
  • [40] IterDE: An Iterative Knowledge Distillation Framework for Knowledge Graph Embeddings
    Liu, Jiajun
    Wang, Peng
    Shang, Ziyu
    Wu, Chenxiao
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 4, 2023, : 4488 - 4496