Evolutionary Recurrent Neural Architecture Search

被引:2
|
作者
Tian, Shuo [1 ]
Hu, Kai [1 ]
Guo, Shasha [1 ]
Li, Shiming [1 ]
Wang, Lei [1 ]
Xu, Weixia [1 ]
机构
[1] Natl Univ Def Technol, Coll Comp Sci & Technol, Changsha 410000, Peoples R China
关键词
Computer architecture; Sociology; Statistics; Microprocessors; Manuals; Training; Computational modeling; Deep learning; evolution algorithm; neural architecture search (NAS); parameter sharing;
D O I
10.1109/LES.2020.3005753
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has promoted remarkable progress in various tasks while the effort devoted to these hand-crafting neural networks has motivated so-called neural architecture search (NAS) to discover them automatically. Recent aging evolution (AE) automatic search algorithm turns to discard the oldest model in population and finds image classifiers beyond manual design. However, it achieves a low speed of convergence. A nonaging evolution (NAE) algorithm tends to neglect the worst architecture in population to accelerate the search process whereas it obtains a lower performance compared with AE. To address this issue, in this letter, we propose to use an optimized evolution algorithm for recurrent NAS (EvoRNAS) by setting a probability epsilon to remove the worst or oldest model in population alternatively, which can balance the performance and time length. Besides, parameter sharing mechanism is introduced in our approach due to the heavy cost of evaluating the candidate models in both AE and NAE. Furthermore, we train the sharing parameters only once instead of many epochs like ENAS, which makes the evaluation of candidate models faster. On Penn Treebank, we first explore different epsilon in EvoRNAS and find the best value suited for the learning task, which is also better than AE and NAE. Second, the optimal cell found by EvoRNAS can achieve state-of-the-art performance within only 0.6 GPU hours, which is 20 x and 40 x faster than ENAS and DARTS. Moreover, the transferability of the learned architecture to WikiText-2 also shows strong performance compared with ENAS or DARTS.
引用
收藏
页码:110 / 113
页数:4
相关论文
共 50 条
  • [31] EvoAAA: An evolutionary methodology for automated neural autoencoder architecture search
    Charte, Francisco
    Rivera, Antonio J.
    Martinez, Francisco
    del Jesus, Maria J.
    [J]. INTEGRATED COMPUTER-AIDED ENGINEERING, 2020, 27 (03) : 211 - 231
  • [32] Efficient evolutionary neural architecture search by modular inheritable crossover
    He, Cheng
    Tan, Hao
    Huang, Shihua
    Cheng, Ran
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2021, 64
  • [33] PRE-NAS: Evolutionary Neural Architecture Search With Predictor
    Peng, Yameng
    Song, Andy
    Ciesielski, Vic
    Fayek, Haytham M. M.
    Chang, Xiaojun
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2023, 27 (01) : 26 - 36
  • [34] Real-Time Federated Evolutionary Neural Architecture Search
    Zhu, Hangyu
    Jin, Yaochu
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2022, 26 (02) : 364 - 378
  • [35] Efficient Self-learning Evolutionary Neural Architecture Search
    Qiu, Zhengzhong
    Bi, Wei
    Xu, Dong
    Guo, Hua
    Ge, Hongwei
    Liang, Yanchun
    Lee, Heow Pueh
    Wu, Chunguo
    [J]. APPLIED SOFT COMPUTING, 2023, 146
  • [36] Evolutionary Neural Architecture Search for Multivariate Time Series Forecasting
    Liang, Zixuan
    Sun, Yanan
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [37] Neural Architecture Search Based on Evolutionary Algorithms with Fitness Approximation
    Pan, Chao
    Yao, Xin
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [38] Evolutionary neural architecture search for remaining useful life prediction
    Mo, Hyunho
    Custode, Leonardo Lucio
    Iacca, Giovanni
    [J]. APPLIED SOFT COMPUTING, 2021, 108
  • [39] Accelerating Evolutionary Neural Architecture Search via Multifidelity Evaluation
    Yang, Shangshang
    Tian, Ye
    Xiang, Xiaoshu
    Peng, Shichen
    Zhang, Xingyi
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (04) : 1778 - 1792
  • [40] Evolutionary Design of Recurrent Neural Network Architecture for Human Activity Recognition
    Viswambaran, Ramya Anasseriyil
    Chen, Gang
    Xue, Bing
    Nekooei, Mohammad
    [J]. 2019 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2019, : 554 - 561