Neural Architecture Search for Computation Offloading of DNNs from Mobile Devices to the Edge Server

被引:0
|
作者
Lee, KyungChae [1 ]
Le Vu Linh [1 ]
Kim, Heejae [1 ]
Youn, Chan-Hyun [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Sch Elect Engn, Daejeon, South Korea
关键词
computational offloading; deep neural network inference; edge computing; neural architecture search;
D O I
10.1109/ICTC52510.2021.9621012
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the rapid development of modern deep learning technology, deep neural network (DNN)-based mobile applications have also been considered for various areas. However, since mobile devices are not optimized to run the DNN applications due to their limit of computational resources, several computation offloading-based approaches have been introduced to overcome the issue; for DNN models, it was reported that, their elaborate partitioning, which allows that input samples are partially executed on mobile devices and then the edge server processes the rest of the execution, can be effective in improving runtime performance. In addition, to improve communication-efficiency in the offloading scenario, there have been also studies to reduce transmitted data from a mobile device and the edge server by leveraging model compression. However, the existing approaches have the root limitation that the performance eventually depend on that of the architecture of original DNN models. To overcome this, we propose a novel neural architecture search (NAS) method to consider the computation offloading cases. On the top of the existing NAS approaches, we additionally introduce resource and channel selection mask. The resource selection mask effectively divides the operations in the target model into those for a mobile device and the edge server; the channel selection mask allows to transmit only selected channels to the edge server without the reduction of task performance (e.g., accuracy). Based on the two additional masks, for the NAS procedure we introduce a new loss function to take into account end-to-end inference time as well as the task performance which is the original goal of NAS. In the evaluation, the proposed method is compared to existing approaches; we see from the experimental results that our method outperforms both the previous NAS and pruning-based model partitioning approaches.
引用
收藏
页码:134 / 139
页数:6
相关论文
共 50 条
  • [1] INTRODUCTION TO THE COMPUTATION OFFLOADING FROM MOBILE DEVICES TO THE EDGE OF MOBILE NETWORK
    Dolezal, Jakub
    Zeman, Tomas
    [J]. ADVANCES IN ELECTRICAL AND ELECTRONIC ENGINEERING, 2019, 17 (04) : 413 - 422
  • [2] Mobile Edge Computing: A Survey on Architecture and Computation Offloading
    Mach, Pavel
    Becvar, Zdenek
    [J]. IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2017, 19 (03): : 1628 - 1656
  • [3] MVR: an Architecture for Computation Offloading in Mobile Edge Computing
    Wei, Xiaojuan
    Wang, Shangguang
    Zhou, Ao
    Xu, Jinliang
    Su, Sen
    Kumar, Sathish
    Yang, Fangchun
    [J]. 2017 IEEE 1ST INTERNATIONAL CONFERENCE ON EDGE COMPUTING (IEEE EDGE), 2017, : 232 - 235
  • [4] Computation Offloading to a Mobile Edge Computing Server with Delay and Energy Constraints
    Hmimz, Youssef
    El Ghmary, Mohamed
    Chanyour, Tarik
    Cherkaoui Malki, Mohammed Oucamah
    [J]. 2019 INTERNATIONAL CONFERENCE ON WIRELESS TECHNOLOGIES, EMBEDDED AND INTELLIGENT SYSTEMS (WITS), 2019,
  • [5] DNNs Based Computation Offloading for LEO Satellite Edge Computing
    Wu, Jian
    Jia, Min
    Zhang, Liang
    Guo, Qing
    [J]. ELECTRONICS, 2022, 11 (24)
  • [6] Cross-Server Computation Offloading for Multi-Task Mobile Edge Computing
    Shi, Yongpeng
    Xia, Yujie
    Gao, Ya
    [J]. INFORMATION, 2020, 11 (02)
  • [7] Cooling-Aware Optimization of Edge Server Configuration and Edge Computation Offloading for Wirelessly Powered Devices
    Chen, Xiaojing
    Lu, Zhouyu
    Ni, Wei
    Wang, Xin
    Wang, Feng
    Zhang, Shunqing
    Xu, Shugong
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (05) : 5043 - 5056
  • [8] IONN: Incremental Offloading of Neural Network Computations from Mobile Devices to Edge Servers
    Jeong, Hyuk-Jin
    Lee, Hyeon-Jae
    Shin, Chang Hyun
    Moon, Soo-Mook
    [J]. PROCEEDINGS OF THE 2018 ACM SYMPOSIUM ON CLOUD COMPUTING (SOCC '18), 2018, : 401 - 411
  • [9] Dynamic Computation Offloading for Mobile-Edge Computing With Energy Harvesting Devices
    Mao, Yuyi
    Zhang, Jun
    Letaief, Khaled B.
    [J]. IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2016, 34 (12) : 3590 - 3605
  • [10] Dynamic Edge Server Placement for Computation Offloading in Vehicular Edge Computing
    Nakrani, Dhruv
    Khuman, Jayesh
    Yadav, Ram Narayan
    [J]. 2023 INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN, 2023, : 45 - 50