Performance Prediction for Convolutional Neural Networks on Edge GPUs

被引:11
|
作者
Bouzidi, Halima [1 ]
Ouarnoughi, Hamza [1 ]
Niar, Smail [1 ]
El Cadi, Abdessamad Ait [1 ]
机构
[1] Univ Polytech Hauts de France, LAMIH, CNRS, Valenciennes, France
关键词
Performance Modeling; CNN; Edge GPU; Multiple Linear Regression; Multi-Layer Perceptrons; Support Vector Machine; Random Forest; XGBoost;
D O I
10.1145/3457388.3458666
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Edge computing is increasingly used for Artificial Intelligence (AI) purposes to meet latency, privacy, and energy challenges. Convolutional Neural networks (CNN) are more frequently deployed on Edge devices for several applications. However, due to their constrained computing resources and energy budget, Edge devices struggle to meet CNN's latency requirements while maintaining good accuracy. It is, therefore, crucial to choose the CNN with the best accuracy and latency trade-off while respecting hardware constraints. This paper presents and compares five of the widely used Machine Learning (ML) based approaches to predict CNN's inference execution time on Edge GPUs. For these 5 methods, in addition to their prediction accuracy, we also explore the time needed for their training and their hyperparameters' tuning. Finally, we compare times to run the prediction models on different platforms. The use of these methods will highly facilitate design space exploration by quickly providing the best CNN on a target Edge GPU. Experimental results show that XGBoost provides an interesting average prediction error even for unexplored and unseen CNN architectures. Random Forest depicts comparable accuracy but needs more effort and time to be trained. The other 3 approaches (OLS, MLP, and SVR) are less accurate for CNN performance estimation.
引用
收藏
页码:54 / 62
页数:9
相关论文
共 50 条
  • [1] High-performance Architecture Aware Sparse Convolutional Neural Networks for GPUs
    Xiang, Lizhi
    Sadayappan, P.
    Sukumaran-Rajam, Aravind
    [J]. PROCEEDINGS OF THE 2022 31ST INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT 2022, 2022, : 265 - 278
  • [2] Analyzing and Increasing the Reliability of Convolutional Neural Networks on GPUs
    dos Santos, Fernando Fernandes
    Pimenta, Pedro Foletto
    Lunardi, Caio
    Draghetti, Lucas
    Carro, Luigi
    Kaeli, David
    Rech, Paolo
    [J]. IEEE TRANSACTIONS ON RELIABILITY, 2019, 68 (02) : 663 - 677
  • [3] Binarized Convolutional Neural Networks for Efficient Inference on GPUs
    Khan, Mir
    Huttunen, Heikki
    Boutellier, Jani
    [J]. 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 682 - 686
  • [4] Prediction of Froth Flotation Performance Using Convolutional Neural Networks
    Jahedsaravani, A.
    Massinaei, M.
    Zarie, M.
    [J]. MINING METALLURGY & EXPLORATION, 2023, 40 (03) : 923 - 937
  • [5] Prediction of Froth Flotation Performance Using Convolutional Neural Networks
    A. Jahedsaravani
    M. Massinaei
    M. Zarie
    [J]. Mining, Metallurgy & Exploration, 2023, 40 : 923 - 937
  • [6] Optimizing Memory Efficiency for Deep Convolutional Neural Networks on GPUs
    Li, Chao
    Yang, Yi
    Feng, Min
    Chakradhar, Srimat
    Zhou, Huiyang
    [J]. SC '16: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2016, : 633 - 644
  • [7] Brief Announcement: On the Limits of Parallelizing Convolutional Neural Networks on GPUs
    Pourghassemi, Behnam
    Zhang, Chenghao
    Lee, Joo Hwan
    Chandramowlishwaran, Aparna
    [J]. PROCEEDINGS OF THE 32ND ACM SYMPOSIUM ON PARALLELISM IN ALGORITHMS AND ARCHITECTURES (SPAA '20), 2020, : 567 - 569
  • [8] Revisiting Edge Detection in Convolutional Neural Networks
    Minh Le
    Kayal, Subhradeep
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] Evaluating the Energy Efficiency of Deep Convolutional Neural Networks on CPUs and GPUs
    Li, Da
    Chen, Xinbo
    Becchi, Michela
    Zong, Ziliang
    [J]. PROCEEDINGS OF 2016 IEEE INTERNATIONAL CONFERENCES ON BIG DATA AND CLOUD COMPUTING (BDCLOUD 2016) SOCIAL COMPUTING AND NETWORKING (SOCIALCOM 2016) SUSTAINABLE COMPUTING AND COMMUNICATIONS (SUSTAINCOM 2016) (BDCLOUD-SOCIALCOM-SUSTAINCOM 2016), 2016, : 477 - 484
  • [10] Performance of Training Sparse Deep Neural Networks on GPUs
    Wang, Jianzong
    Huang, Zhangcheng
    Kong, Lingwei
    Xiao, Jing
    Wang, Pengyu
    Zhang, Lu
    Li, Chao
    [J]. 2019 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2019,