SLAPP: Subgraph-level attention-based performance prediction for deep learning models

被引:1
|
作者
Wang, Zhenyi [1 ,2 ]
Yang, Pengfei [1 ]
Hu, Linwei [1 ,2 ]
Zhang, Bowen [1 ,2 ]
Lin, Chengmin [1 ,2 ]
Lv, Wenkai [1 ,2 ]
Wang, Quan [1 ,2 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian 710071, Peoples R China
[2] Key Lab Smart Human Comp Interact & Wearable Techn, Xian 710071, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep Learning (DL); Graph neural networks (GNNs); Performance prediction; Computation graph optimization; Attention mechanisms; NEURAL-NETWORK;
D O I
10.1016/j.neunet.2023.11.043
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The intricacy of the Deep Learning (DL) landscape, brimming with a variety of models, applications, and platforms, poses considerable challenges for the optimal design, optimization, or selection of suitable DL models. One promising avenue to address this challenge is the development of accurate performance prediction methods. However, existing methods reveal critical limitations. Operator-level methods, proficient at predicting the performance of individual operators, often neglect broader graph features, which results in inaccuracies in full network performance predictions. On the contrary, graph-level methods excel in overall network prediction by leveraging these graph features but lack the ability to predict the performance of individual operators. To bridge these gaps, we propose SLAPP, a novel subgraph-level performance prediction method. Central to SLAPP is an innovative variant of Graph Neural Networks (GNNs) that we developed, named the Edge Aware Graph Attention Network (EAGAT). This specially designed GNN enables superior encoding of both node and edge features. Through this approach, SLAPP effectively captures both graph and operator features, thereby providing precise performance predictions for individual operators and entire networks. Moreover, we introduce a mixed loss design with dynamic weight adjustment to reconcile the predictive accuracy between individual operators and entire networks. In our experimental evaluation, SLAPP consistently outperforms traditional approaches in prediction accuracy, including the ability to handle unseen models effectively. Moreover, when compared to existing research, our method demonstrates a superior predictive performance across multiple DL models.
引用
收藏
页码:285 / 297
页数:13
相关论文
共 50 条
  • [21] Enhancing intra-aural disease classification with attention-based deep learning models
    Furkancan Demircan
    Murat Ekinci
    Zafer Cömert
    Neural Computing and Applications, 2025, 37 (9) : 6601 - 6616
  • [22] Attention-based Deep Learning for Network Intrusion Detection
    Guo, Naiwang
    Tian, Yingjie
    Li, Fan
    Yang, Hongshan
    2020 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING AND ARTIFICIAL INTELLIGENCE, 2020, 11584
  • [23] On the Instability of Softmax Attention-Based Deep Learning Models in Side-Channel Analysis
    Hajra, Suvadeep
    Alam, Manaar
    Saha, Sayandeep
    Picek, Stjepan
    Mukhopadhyay, Debdeep
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 514 - 528
  • [24] Attention-Based Light Weight Deep Learning Models for Early Potato Disease Detection
    Kasana, Singara Singh
    Rathore, Ajayraj Singh
    APPLIED SCIENCES-BASEL, 2024, 14 (17):
  • [25] Attention-based learning
    Kasderidis, S
    Taylor, JG
    2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 525 - 530
  • [26] Deep Attention-Based Classification Network for Robust Depth Prediction
    Li, Ruibo
    Xian, Ke
    Shen, Chunhua
    Cao, Zhiguo
    Lu, Hao
    Hang, Lingxiao
    COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 663 - 678
  • [27] Mitigating Biases in Student Performance Prediction via Attention-Based Personalized Federated Learning
    Chu, Yun-Wei
    Hosseinalipour, Seyyedali
    Tenorio, Elizabeth
    Cruz, Laura
    Douglas, Kerrie
    Lan, Andrew
    Brinton, Christopher
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3033 - 3042
  • [28] Spatio-Temporal Attention-Based Deep Learning Framework for Mesoscale Eddy Trajectory Prediction
    Wang, Xuegong
    Li, Chong
    Wang, Xinning
    Tan, Lining
    Wu, Jin
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 : 3853 - 3867
  • [29] DMSS: An Attention-Based Deep Learning Model for High-Quality Mass Spectrometry Prediction
    Ren, Yihui
    Wang, Yu
    Han, Wenkai
    Huang, Yikang
    Hou, Xiaoyang
    Zhang, Chunming
    Bu, Dongbo
    Gao, Xin
    Sun, Shiwei
    BIG DATA MINING AND ANALYTICS, 2024, 7 (03): : 577 - 589
  • [30] Ensemble CNN Attention-Based BiLSTM Deep Learning Architecture for Multivariate Cloud Workload Prediction
    Kaim, Ananya
    Singh, Surjit
    Patel, Yashwant Singh
    PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING AND NETWORKING, ICDCN 2023, 2023, : 342 - 348