Multihardware Adaptive Latency Prediction for Neural Architecture Search

被引:0
|
作者
Lin, Chengmin [1 ]
Yang, Pengfei [1 ]
Wang, Quan [1 ]
Guo, Yitong [1 ]
Wang, Zhenyi [1 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian 710126, Peoples R China
来源
IEEE INTERNET OF THINGS JOURNAL | 2025年 / 12卷 / 03期
关键词
Hardware; Predictive models; Adaptation models; Training; Accuracy; Network architecture; Computer architecture; Optimization; Performance evaluation; Data models; Dynamic sample allocation; few-shot learning; hardware-aware; latency predictor; neural architecture search (NAS); representative sample sampling; NETWORKS;
D O I
10.1109/JIOT.2024.3480990
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In hardware-aware neural architecture search (NAS), accurately assessing a model's inference efficiency is crucial for search optimization. Traditional approaches, which measure numerous samples to train proxy models, are impractical across varied platforms due to the extensive resources needed to remeasure and rebuild models for each platform. To address this challenge, we propose a multihardware-aware NAS method that enhances the generalizability of proxy models across different platforms while reducing the required sample size. Our method introduces a multihardware adaptive latency prediction (MHLP) model that leverages one-hot encoding for hardware parameters and multihead attention mechanisms to effectively capture the intricate interplay between hardware attributes and network architecture features. Additionally, we implement a two-stage sampling mechanism based on probability density weighting to ensure the representativeness and diversity of the sample set. By adopting a dynamic sample allocation mechanism, our method can adjust the adaptive sample size according to the initial model state, providing stronger data support for devices with significant deviations. Evaluations on NAS benchmarks demonstrate the MHLP predictor's excellent generalization accuracy using only 10 samples, guiding the NAS search process to identify optimal network architectures.
引用
收藏
页码:3385 / 3398
页数:14
相关论文
共 50 条
  • [41] Diversity in Neural Architecture Search
    Hu, Wenzheng
    Li, Mingyang
    Yuan, Changhe
    Zhang, Changshui
    Wang, Jianqiang
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [42] Surrogate-Assisted Evolutionary Neural Architecture Search with Isomorphic Training and Prediction
    Jiang, Pengcheng
    Xue, Yu
    Neri, Ferrante
    Wahib, Mohamed
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14863 : 191 - 203
  • [43] A review of neural architecture search
    Baymurzina, Dilyara
    Golikov, Eugene
    Burtsev, Mikhail
    NEUROCOMPUTING, 2022, 474 : 82 - 93
  • [44] Disentangled Neural Architecture Search
    Zheng, Xinyue
    Wang, Peng
    Wang, Qigang
    Shi, Zhongchao
    Fan, Jianping
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [45] An evolutionary neural architecture search method based on performance prediction and weight inheritance
    Yuan G.
    Xue B.
    Zhang M.
    Information Sciences, 2024, 667
  • [46] MicroNAS for memory and latency constrained hardware aware neural architecture search in time series classification on microcontrollers
    King, Tobias
    Zhou, Yexu
    Roeddiger, Tobias
    Beigl, Michael
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [47] Evolving Search Space for Neural Architecture Search
    Ci, Yuanzheng
    Lin, Chen
    Sun, Ming
    Chen, Boyu
    Zhang, Hongwen
    Ouyang, Wanli
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6639 - 6649
  • [48] Random Search and Reproducibility for Neural Architecture Search
    Li, Liam
    Talwalkar, Ameet
    35TH UNCERTAINTY IN ARTIFICIAL INTELLIGENCE CONFERENCE (UAI 2019), 2020, 115 : 367 - 377
  • [49] Effective, Efficient and Robust Neural Architecture Search Effective, Efficient and Robust Neural Architecture Search
    Yue, Zhixiong
    Lin, Baijiong
    Zhang, Yu
    Liang, Christy
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [50] Neural Architecture Search for Spiking Neural Networks
    Kim, Youngeun
    Li, Yuhang
    Park, Hyoungseob
    Venkatesha, Yeshwanth
    Panda, Priyadarshini
    COMPUTER VISION, ECCV 2022, PT XXIV, 2022, 13684 : 36 - 56