Implication of Optimizing NPU Dataflows on Neural Architecture Search for Mobile Devices

被引:0
|
作者
Lee, Jooyeon [1 ]
Park, Junsang [1 ]
Lee, Seunghyun [1 ]
Kung, Jaeha [1 ]
机构
[1] Daegu Gyeongbuk Inst Sci & Technol DGIST, Daegu 42988, South Korea
关键词
Dataflow optimization; neural networks; neural architecture search; neural processing unit;
D O I
10.1145/3513085
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Recent advances in deep learning have made it possible to implement artificial intelligence in mobile devices. Many studies have put a lot of effort into developing lightweight deep learning models optimized for mobile devices. To overcome the performance limitations of manually designed deep learning models, an automated search algorithm, called neural architecture search (NAS), has been proposed. However, studies on the effect of hardware architecture of the mobile device on the performance of NAS have been less explored. In this article, we show the importance of optimizing a hardware architecture, namely, NPU dataflow, when searching for a more accurate yet fast deep learning model. To do so, we first implement an optimization framework, named FlowOptimizer, for generating a best possible NPU dataflow for a given deep learning operator. Then, we utilize this framework during the latency-aware NAS to find the model with the highest accuracy satisfying the latency constraint. As a result, we show that the searched model with FlowOptimizer outperforms the performance by 87.1% and 92.3% on average compared to the searched model with NVDLA and Eyeriss, respectively, with better accuracy on a proxy dataset. We also show that the searched model can be transferred to a larger model to classify a more complex image dataset, i.e., ImageNet, achieving 0.2%/5.4% higher Top-1/Top-5 accuracy compared to MobileNetV2-1.0 with 3.6x lower latency.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] Progressive Neural Architecture Search
    Liu, Chenxi
    Zoph, Barret
    Neumann, Maxim
    Shlens, Jonathon
    Hua, Wei
    Li, Li-Jia
    Li Fei-Fei
    Yuille, Alan
    Huang, Jonathan
    Murphy, Kevin
    COMPUTER VISION - ECCV 2018, PT I, 2018, 11205 : 19 - 35
  • [42] Advances in neural architecture search
    Wang, Xin
    Zhu, Wenwu
    NATIONAL SCIENCE REVIEW, 2024, 11 (08)
  • [43] Personalized Neural Architecture Search
    Kulbach, Cedric
    Thoma, Steffen
    21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 581 - 590
  • [44] Binarized Neural Architecture Search
    Chen, Hanlin
    Zhuo, Li'an
    Zhang, Baochang
    Zheng, Xiawu
    Liu, Jianzhuang
    Doermann, David
    Ji, Rongrong
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 10526 - 10533
  • [45] Neural Architecture Search: A Survey
    Elsken, Thomas
    Metzen, Jan Hendrik
    Hutter, Frank
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [46] Balanced neural architecture search
    Li, Yangyang
    Liu, Guanlong
    Zhao, Peixiang
    Shang, Ronghua
    Jiao, Licheng
    NEUROCOMPUTING, 2024, 594
  • [47] Hypergraph Neural Architecture Search
    Lin, Wei
    Peng, Xu
    Yu, Zhengtao
    Jin, Taisong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 12, 2024, : 13837 - 13845
  • [48] Diversity in Neural Architecture Search
    Hu, Wenzheng
    Li, Mingyang
    Yuan, Changhe
    Zhang, Changshui
    Wang, Jianqiang
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [49] A review of neural architecture search
    Baymurzina, Dilyara
    Golikov, Eugene
    Burtsev, Mikhail
    NEUROCOMPUTING, 2022, 474 : 82 - 93
  • [50] Disentangled Neural Architecture Search
    Zheng, Xinyue
    Wang, Peng
    Wang, Qigang
    Shi, Zhongchao
    Fan, Jianping
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,