Asynchronous evolution of deep neural network architectures

被引:0
|
作者
Liang, Jason [1 ]
Shahrzad, Hormoz [1 ]
Miikkulainen, Risto [1 ,2 ]
机构
[1] Cognizant AI Labs, Teaneck, NJ 07666 USA
[2] Univ Texas Austin, Austin, TX 78712 USA
关键词
Evolutionary computation; Parallelization; Asynchronous evolution; Sorting networks; Multiplexer design; Neural architecture search; Neuroevolution; NEUROEVOLUTION;
D O I
10.1016/j.asoc.2023.111209
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many evolutionary algorithms (EAs) take advantage of parallel evaluation of candidates. However, if evaluation times vary significantly, many worker nodes (i.e., compute clients) are idle much of the time, waiting for the next generation to be created. Evolutionary neural architecture search (ENAS), a class of EAs that optimizes the architecture and hyperparameters of deep neural networks, is particularly vulnerable to this issue. This paper proposes a generic asynchronous evaluation strategy (AES) that is then adapted to work with ENAS. AES increases throughput by maintaining a queue of up to K individuals ready to be sent to the workers for evaluation and proceeding to the next generation as soon as M << K individuals have been evaluated. A suitable value for M is determined experimentally, balancing diversity and efficiency. To showcase the generality and power of AES, it was first evaluated in eight-line sorting network design (a single-population optimization task with limited evaluation-time variability), achieving an over two-fold speedup. Next, it was evaluated in 11-bit multiplexer design (a single-population discovery task with extended variability), where a 14-fold speedup was observed. It was then scaled up to ENAS for image captioning (a multi-population openended-optimization task), resulting in an over two-fold speedup. In all problems, a multifold performance improvement was observed, suggesting that AES is a promising method for parallelizing the evolution of complex systems with long and variable evaluation times, such as those in ENAS.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Neural network with deep learning architectures
    Patel, Hima
    Thakkar, Amit
    Pandya, Mrudang
    Makwana, Kamlesh
    [J]. JOURNAL OF INFORMATION & OPTIMIZATION SCIENCES, 2018, 39 (01): : 31 - 38
  • [2] Deep Neural Network Architectures for Modulation Classification
    Liu, Xiaoyu
    Yang, Diyu
    El Gamal, Aly
    [J]. 2017 FIFTY-FIRST ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2017, : 915 - 919
  • [3] A survey of deep neural network architectures and their applications
    Liu, Weibo
    Wang, Zidong
    Liu, Xiaohui
    Zeng, Nianyin
    Liu, Yurong
    Alsaadi, Fuad E.
    [J]. NEUROCOMPUTING, 2017, 234 : 11 - 26
  • [4] Optimizing Deep Neural Network Architectures: an overview
    Bouzar-Benlabiod, Lydia
    Rubin, Stuart H.
    Benaida, Amel
    [J]. 2021 IEEE 22ND INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE (IRI 2021), 2021, : 25 - 32
  • [5] AUTOMATED HARDENING OF DEEP NEURAL NETWORK ARCHITECTURES
    Beyer, Michael
    Schorn, Christoph
    Fabarisov, Tagir
    Morozov, Andrey
    Janschek, Klaus
    [J]. PROCEEDINGS OF ASME 2021 INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION (IMECE2021), VOL 13, 2021,
  • [6] Evolution of Neural Network Architectures for Speech Recognition
    Bourlard, Herve
    [J]. 19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 1767 - 1767
  • [7] Design and evolution of modular neural network architectures
    Happel, Bart L.M.
    Murre, Jacob M.J.
    [J]. 1600, Pergamon Press Inc, Tarrytown, NY, United States (07): : 6 - 7
  • [8] Dedicated Deep Neural Network Architectures and Methods for Their Training
    Rozycki, P.
    Kolbusz, J.
    Wilamowski, B. M.
    [J]. INES 2015 - IEEE 19TH INTERNATIONAL CONFERENCE ON INTELLIGENT ENGINEERING SYSTEMS, 2015, : 73 - 78
  • [9] Efficient Hardware Architectures for Deep Convolutional Neural Network
    Wang, Jichen
    Lin, Jun
    Wang, Zhongfeng
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (06) : 1941 - 1953
  • [10] Benchmark Analysis of Representative Deep Neural Network Architectures
    Bianco, Simone
    Cadene, Remi
    Celona, Luigi
    Napoletano, Paolo
    [J]. IEEE ACCESS, 2018, 6 : 64270 - 64277