DGL: Device Generic Latency Model for Neural Architecture Search on Mobile Devices

被引:0
|
作者
Wang, Qinsi [1 ]
Zhang, Sihai [2 ,3 ]
机构
[1] Univ Sci & Technol China, Dept Elect Sci & Technol, Hefei 230026, Anhui, Peoples R China
[2] Chinese Acad Sci, Key Lab Wireless Opt Commun, Beijing 100045, Peoples R China
[3] Univ Sci & Technol China, Sch Microelect, Hefei 230026, Anhui, Peoples R China
关键词
Predictive models; Training; Mobile handsets; Hardware; Costs; Computer architecture; Analytical models; Neural architecture search (NAS); processor interval analysis; latency prediction; mobile devices;
D O I
10.1109/TMC.2023.3244170
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The low-cost Neural Architecture Search (NAS) for lightweight networks working on massive mobile devices is essential for fast-developing ICT technology. Current NAS work can not search on unseen devices without latency sampling, which is a big obstacle to the implementation of NAS on mobile devices. In this paper, we overcome this challenge by proposing the Device Generic Latency (DGL) model. By absorbing processor modeling technology, the proposed DGL formula maps the parameters in the interval theory to the seven static configuration parameters of the device. And to make the formula more practical, we refine it to low-cost form by decreasing the number of configuration parameters to four. Then based on this formula, the DGL model is proposed which introduces the network parameters predictor and accuracy predictor to work with the DGL formula to predict the network latency. We propose the DGL-based NAS framework to enable fast searches without latency sampling. Extensive experiments results validate that the DGL model can achieve more accurate latency predictions than existing NAS latency predictors on unseen mobile devices. When configured with current state-of-the-art predictors, DGL-based NAS can search for architectures with higher accuracy that meet the latency limit than other NAS implementations, while using less training time and prediction time. Our work shed light on how to adopt domain knowledge into NAS topic and play important role in low-cost NAS on mobile devices.
引用
收藏
页码:1954 / 1967
页数:14
相关论文
共 50 条
  • [31] Reducing Power Consumption and Latency in Mobile Devices Using an Event Stream Model
    Marz, Stephen
    Vander Zanden, Brad
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2016, 16 (01)
  • [32] A generic deep learning architecture optimization method for edge device based on start-up latency reduction
    Li, Qi
    Li, Hengyi
    Meng, Lin
    JOURNAL OF REAL-TIME IMAGE PROCESSING, 2024, 21 (04)
  • [33] A generic LSTM neural network architecture to infer heterogeneous model transformations
    Loli Burgueño
    Jordi Cabot
    Shuai Li
    Sébastien Gérard
    Software and Systems Modeling, 2022, 21 : 139 - 156
  • [34] A generic LSTM neural network architecture to infer heterogeneous model transformations
    Burgueno, Loli
    Cabot, Jordi
    Li, Shuai
    Gerard, Sebastien
    SOFTWARE AND SYSTEMS MODELING, 2022, 21 (01): : 139 - 156
  • [35] Enhancing Neural Architecture Search With Multiple Hardware Constraints for Deep Learning Model Deployment on Tiny IoT Devices
    Burrello, Alessio
    Risso, Matteo
    Motetti, Beatrice Alessandra
    Macii, Enrico
    Benini, Luca
    Pagliari, Daniele Jahier
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2024, 12 (03) : 780 - 794
  • [36] HGNAS: Hardware-Aware Graph Neural Architecture Search for Edge Devices
    Zhou, Ao
    Yang, Jianlei
    Qi, Yingjie
    Qiao, Tong
    Shi, Yumeng
    Duan, Cenlin
    Zhao, Weisheng
    Hu, Chunming
    IEEE Transactions on Computers, 2024, 73 (12) : 2693 - 2707
  • [37] TAS: Ternarized Neural Architecture Search for Resource-Constrained Edge Devices
    Loni, Mohammad
    Mousavi, Hamid
    Riazati, Mohammad
    Daneshtalab, Masoud
    Sjodin, Mikael
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 1115 - 1118
  • [38] HyT-NAS: Hybrid Transformers Neural Architecture Search for Edge Devices
    Mecharbat, Lotfi Abdelkrim
    Benmeziane, Hadjer
    Ouarnoughi, Hamza
    Niar, Smail
    PROCEEDINGS 2023 IEEE/ACM INTERNATIONAL WORKSHOP ON COMPILERS, DEPLOYMENT, AND TOOLING FOR EDGE AI, CODAI 2023, 2023, : 41 - 45
  • [39] Privacy-Preserving Neural Architecture Search Across Federated IoT Devices
    Zhang, Chunhui
    Yuan, Xiaoming
    Zhang, Qianyun
    Zhu, Guangxu
    Cheng, Lei
    Zhang, Ning
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1434 - 1438
  • [40] One Proxy Device Is Enough for Hardware-Aware Neural Architecture Search
    Lu, Bingqian
    Yang, Jianyi
    Jiang, Weiwen
    Shi, Yiyu
    Ren, Shaolei
    PROCEEDINGS OF THE ACM ON MEASUREMENT AND ANALYSIS OF COMPUTING SYSTEMS, 2021, 5 (03)