An Empirical Study of Challenges in Converting Deep Learning Models

被引:9
|
作者
Openja, Moses [1 ]
Nikanjam, Amin [1 ]
Yahmed, Ahmed Haj [1 ]
Khomh, Foutse [1 ]
Jiang, Zhen Ming [2 ]
机构
[1] Polytech Montreal, Montreal, PQ, Canada
[2] York Univ, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Empirical; Deep Learning; Converting Trained Models; Deploying ML Models; Robustness;
D O I
10.1109/ICSME55016.2022.00010
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
There is an increase in deploying Deep Learning (DL)-based software systems in real-world applications. Usually, DL models are developed and trained using DL frameworks like TensorFlow and PyTorch. Each framework has its own internal mechanisms/formats to represent and train DL models (deep neural networks), and usually those formats cannot be recognized by other frameworks. Moreover, trained models are usually deployed in environments different from where they were developed. To solve the interoperability issue and make DL models compatible with different frameworks/environments, some exchange formats are introduced for DL models, like ONNX and CoreML. However, ONNX and CoreML were never empirically evaluated by the community to reveal their prediction accuracy, performance, and robustness after conversion. Poor accuracy or non-robust behavior of converted models may lead to poor quality of deployed DL-based software systems. We conduct, in this paper, the first empirical study to assess ONNX and CoreML for converting trained DL models. In our systematic approach, two popular DL frameworks, Keras and PyTorch, are used to train five widely used DL models on three popular datasets. The trained models are then converted to ONNX and CoreML and transferred to two runtime environments designated for such formats, to be evaluated. We investigate the prediction accuracy before and after conversion. Our results unveil that the prediction accuracy of converted models are at the same level of originals. The performance (time cost and memory consumption) of converted models are studied as well. The size of models are reduced after conversion, which can result in optimized DL-based software deployment. We also study the adversarial robustness of converted models to make sure about the robustness of deployed DL-based software. Leveraging the state-of-the-art adversarial attack approaches, converted models are generally assessed robust at the same level of originals. However, obtained results show that CoreML models are more vulnerable to adversarial attacks compared to ONNX. The general message of our findings is that DL developers should be cautious on the deployment of converted models that may 1) perform poorly while switching from one framework to another, 2) have challenges in robust deployment, or 3) run slowly, leading to poor quality of deployed DL-based software, including DL-based software maintenance tasks, like bug prediction.
引用
收藏
页码:13 / 23
页数:11
相关论文
共 50 条
  • [41] Challenges to Student Interdisciplinary Learning Effectiveness: An Empirical Case Study
    Xu, Cong
    Wu, Chih-Fu
    Xu, Dan-Dan
    Lu, Wen-Qian
    Wang, Kai-Yi
    [J]. JOURNAL OF INTELLIGENCE, 2022, 10 (04)
  • [42] Deep learning based automatic modulation recognition: Models, datasets, and challenges
    Zhang, Fuxin
    Luo, Chunbo
    Xu, Jialang
    Luo, Yang
    Zheng, Fu-Chun
    [J]. DIGITAL SIGNAL PROCESSING, 2022, 129
  • [43] Deep Learning for Source Code Modeling and Generation: Models, Applications, and Challenges
    Le, Triet H. M.
    Chen, Hao
    Babar, Muhammad Ali
    [J]. ACM COMPUTING SURVEYS, 2020, 53 (03)
  • [44] Deep Learning Models for Medical Image Analysis: Challenges and Future Directions
    Agrawal, R. K.
    Juneja, Akanksha
    [J]. BIG DATA ANALYTICS (BDA 2019), 2019, 11932 : 20 - 32
  • [45] Deep learning models for predictive maintenance: a survey, comparison, challenges and prospects
    Oscar Serradilla
    Ekhi Zugasti
    Jon Rodriguez
    Urko Zurutuza
    [J]. Applied Intelligence, 2022, 52 : 10934 - 10964
  • [46] Data management for production quality deep learning models: Challenges and solutions
    Munappy, Aiswarya Raj
    Bosch, Jan
    Olsson, Helena Holmstrom
    Arpteg, Anders
    Brinne, Bjoern
    [J]. JOURNAL OF SYSTEMS AND SOFTWARE, 2022, 191
  • [47] Deep learning models for predictive maintenance: a survey, comparison, challenges and prospects
    Serradilla, Oscar
    Zugasti, Ekhi
    Rodriguez, Jon
    Zurutuza, Urko
    [J]. APPLIED INTELLIGENCE, 2022, 52 (10) : 10934 - 10964
  • [48] A Review of Deep Learning Models for Twitter Sentiment Analysis: Challenges and Opportunities
    Chaudhary, Laxmi
    Girdhar, Nancy
    Sharma, Deepak
    Andreu-Perez, Javier
    Doucet, Antoine
    Renz, Matthias
    [J]. IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (03): : 3550 - 3579
  • [49] Deep Learning on Microcontrollers: A Study on Deployment Costs and Challenges
    Svoboda, Filip
    Fernandez-Marques, Javier
    Liberis, Edgar
    Lane, Nicholas D.
    [J]. PROCEEDINGS OF THE 2022 2ND EUROPEAN WORKSHOP ON MACHINE LEARNING AND SYSTEMS (EUROMLSYS '22), 2022, : 54 - 63
  • [50] Empirical Forecasting Analysis of Bitcoin Prices: A Comparison of Machine Learning, Deep Learning, and Ensemble Learning Models
    Tripathy, Nrusingha
    Hota, Sarbeswara
    Mishra, Debahuti
    Satapathy, Pranati
    Nayak, Subrat Kumar
    [J]. INTERNATIONAL JOURNAL OF ELECTRICAL AND COMPUTER ENGINEERING SYSTEMS, 2024, 15 (01) : 21 - 29