Multihardware Adaptive Latency Prediction for Neural Architecture Search

被引:0
|
作者
Lin, Chengmin [1 ]
Yang, Pengfei [1 ]
Wang, Quan [1 ]
Guo, Yitong [1 ]
Wang, Zhenyi [1 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian 710126, Peoples R China
来源
IEEE INTERNET OF THINGS JOURNAL | 2025年 / 12卷 / 03期
关键词
Hardware; Predictive models; Adaptation models; Training; Accuracy; Network architecture; Computer architecture; Optimization; Performance evaluation; Data models; Dynamic sample allocation; few-shot learning; hardware-aware; latency predictor; neural architecture search (NAS); representative sample sampling; NETWORKS;
D O I
10.1109/JIOT.2024.3480990
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In hardware-aware neural architecture search (NAS), accurately assessing a model's inference efficiency is crucial for search optimization. Traditional approaches, which measure numerous samples to train proxy models, are impractical across varied platforms due to the extensive resources needed to remeasure and rebuild models for each platform. To address this challenge, we propose a multihardware-aware NAS method that enhances the generalizability of proxy models across different platforms while reducing the required sample size. Our method introduces a multihardware adaptive latency prediction (MHLP) model that leverages one-hot encoding for hardware parameters and multihead attention mechanisms to effectively capture the intricate interplay between hardware attributes and network architecture features. Additionally, we implement a two-stage sampling mechanism based on probability density weighting to ensure the representativeness and diversity of the sample set. By adopting a dynamic sample allocation mechanism, our method can adjust the adaptive sample size according to the initial model state, providing stronger data support for devices with significant deviations. Evaluations on NAS benchmarks demonstrate the MHLP predictor's excellent generalization accuracy using only 10 samples, guiding the NAS search process to identify optimal network architectures.
引用
收藏
页码:3385 / 3398
页数:14
相关论文
共 50 条
  • [31] Graph Neural Architecture Search
    Gao, Yang
    Yang, Hong
    Zhang, Peng
    Zhou, Chuan
    Hu, Yue
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1403 - 1409
  • [32] Neural architecture search: A survey
    Elsken, Thomas
    Metzen, Jan Hendrik
    Hutter, Frank
    Journal of Machine Learning Research, 2019, 20
  • [33] Advances in neural architecture search
    Xin Wang
    Wenwu Zhu
    National Science Review, 2024, 11 (08) : 24 - 38
  • [34] Progressive Neural Architecture Search
    Liu, Chenxi
    Zoph, Barret
    Neumann, Maxim
    Shlens, Jonathon
    Hua, Wei
    Li, Li-Jia
    Li Fei-Fei
    Yuille, Alan
    Huang, Jonathan
    Murphy, Kevin
    COMPUTER VISION - ECCV 2018, PT I, 2018, 11205 : 19 - 35
  • [35] Advances in neural architecture search
    Wang, Xin
    Zhu, Wenwu
    NATIONAL SCIENCE REVIEW, 2024, 11 (08)
  • [36] Personalized Neural Architecture Search
    Kulbach, Cedric
    Thoma, Steffen
    21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 581 - 590
  • [37] Binarized Neural Architecture Search
    Chen, Hanlin
    Zhuo, Li'an
    Zhang, Baochang
    Zheng, Xiawu
    Liu, Jianzhuang
    Doermann, David
    Ji, Rongrong
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 10526 - 10533
  • [38] Neural Architecture Search: A Survey
    Elsken, Thomas
    Metzen, Jan Hendrik
    Hutter, Frank
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [39] Balanced neural architecture search
    Li, Yangyang
    Liu, Guanlong
    Zhao, Peixiang
    Shang, Ronghua
    Jiao, Licheng
    NEUROCOMPUTING, 2024, 594
  • [40] Hypergraph Neural Architecture Search
    Lin, Wei
    Peng, Xu
    Yu, Zhengtao
    Jin, Taisong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 12, 2024, : 13837 - 13845