Learning structure perception MLPs on graphs: a layer-wise graph knowledge distillation framework

被引:0
|
作者
Du, Hangyuan [1 ]
Yu, Rong [1 ]
Bai, Liang [2 ,3 ]
Bai, Lu [4 ,5 ]
Wang, Wenjian [2 ]
机构
[1] Shanxi Univ, Sch Comp & Informat Technol, Taiyuan 030006, Shanxi, Peoples R China
[2] Shanxi Univ, Key Lab Computat Intelligence Chinese Informat Pro, Minist Educ, Taiyuan 030006, Shanxi, Peoples R China
[3] Shanxi Univ, Inst Intelligent Informat Proc, Taiyuan 030006, Shanxi, Peoples R China
[4] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
[5] Cent Univ Finance & Econ, Beijing 100875, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph knowledge distillation; Supervision signal; Layer-wise mapping; Structure perception MLPs;
D O I
10.1007/s13042-024-02150-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) are expressive in dealing with graph data. Because of the large storage requirements and the high computational complexity, it is difficult to deploy these cumbersome models in resource-constrained environments. As a representative model compression strategy, knowledge distillation (KD) is introduced into graph analysis research to address this problem. However, there are some crucial challenges in existing graph knowledge distillation algorithms, such as knowledge transfer effectiveness and student model designation. To address these problems, a new graph distillation model is proposed in this paper. Specifically, a layer-wise mapping strategy is designed to distill knowledge for training the student model, in which staged knowledge learned by intermediate layers of teacher GNNs is captured to form supervision signals. And, an adaptive weight mechanism is developed to evaluate the importance of the distilled knowledge. On this basis, a structure perception MLPs is constructed as the student model, which can capture prior information of the input graph from the perspectives of node feature and topology structure. In this way, the proposed model shares the prediction advantage of GNNs and the latency advantage of MLPs. Node classification experiments on five benchmark datasets demonstrate the validity and superiority of our model over baseline algorithms.
引用
收藏
页码:4357 / 4372
页数:16
相关论文
共 50 条
  • [41] Layer-Wise Geometry Aggregation Framework for Lossless LiDAR Point Cloud Compression
    Song, Fei
    Shao, Yiting
    Gao, Wei
    Wang, Haiqiang
    Li, Thomas
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (12) : 4603 - 4616
  • [42] Layer-wise regularized adversarial training using layers sustainability analysis framework
    Khalooei, Mohammad
    Homayounpour, Mohammad Mehdi
    Amirmazlaghani, Maryam
    [J]. NEUROCOMPUTING, 2023, 540
  • [43] Cost-Sensitive Deep Learning with Layer-Wise Cost Estimation
    Chung, Yu-An
    Yang, Shao-Wen
    Lin, Hsuan-Tien
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE (TAAI 2020), 2020, : 108 - 113
  • [44] A deep learning framework for layer-wise porosity prediction in metal powder bed fusion using thermal signatures
    Yuwei Mao
    Hui Lin
    Christina Xuan Yu
    Roger Frye
    Darren Beckett
    Kevin Anderson
    Lars Jacquemetton
    Fred Carter
    Zhangyuan Gao
    Wei-keng Liao
    Alok N. Choudhary
    Kornel Ehmann
    Ankit Agrawal
    [J]. Journal of Intelligent Manufacturing, 2023, 34 : 315 - 329
  • [45] A deep learning framework for layer-wise porosity prediction in metal powder bed fusion using thermal signatures
    Mao, Yuwei
    Lin, Hui
    Yu, Christina Xuan
    Frye, Roger
    Beckett, Darren
    Anderson, Kevin
    Jacquemetton, Lars
    Carter, Fred
    Gao, Zhangyuan
    Liao, Wei-keng
    Choudhary, Alok N.
    Ehmann, Kornel
    Agrawal, Ankit
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 2023, 34 (01) : 315 - 329
  • [46] Fed-LAMB: Layer-wise and Dimension-wise Locally Adaptive Federated Learning
    Karimi, Belhal
    Li, Ping
    Li, Xiaoyun
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1037 - 1046
  • [47] Personalized Federated Learning with Layer-Wise Feature Transformation via Meta-Learning
    Tu, Jingke
    Huang, Jiaming
    Yang, Lei
    Lin, Wanyu
    [J]. ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2024, 18 (04)
  • [48] Graph Structure Aware Contrastive Knowledge Distillation for Incremental Learning in Recommender Systems
    Wang, Yuening
    Zhang, Yingxue
    Coates, Mark
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3518 - 3522
  • [49] Layer-Wise Adaptive Gradient Sparsification for Distributed Deep Learning with Convergence Guarantees
    Shi, Shaohuai
    Tang, Zhenheng
    Wang, Qiang
    Zhao, Kaiyong
    Chu, Xiaowen
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1467 - 1474
  • [50] Investigating Learning in Deep Neural Networks Using Layer-Wise Weight Change
    Agrawal, Ayush Manish
    Tendle, Atharva
    Sikka, Harshvardhan
    Singh, Sahib
    Kayid, Amr
    [J]. INTELLIGENT COMPUTING, VOL 2, 2021, 284 : 678 - 693