Recurrent Neural Architecture Search based on Randomness-Enhanced Tabu Algorithm

被引:0
|
作者
Hu, Kai [1 ]
Tian, Shuo [1 ]
Guo, Shasha [1 ]
Li, Nan [1 ]
Luo, Li [1 ]
Wang, Lei [1 ]
机构
[1] Natl Univ Def Technol, Coll Comp Sci & Technol, Changsha, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
deep learning; neural architecture search; tabu algorithm; weight sharing;
D O I
10.1109/ijcnn48605.2020.9207393
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have achieved highly competitive performance in multiple tasks in recent years. However, discovering state-of-the-art neural network architectures requires substantial effort from human experts. To speed up the process, neural architecture search (NAS) has been proposed to search promising architectures automatically. Nevertheless, the search process of NAS is computing-expensive and time-consuming, which even costs thousands of GPU days. In this paper, to solve the bottleneck, we apply the randomness-enhanced tabu algorithm as a controller to sample candidate architectures, which balances the global exploration and local exploitation for the architectural solutions. In addition, more aggressive weight-sharing strategy is introduced into our method, which significantly reduces the overhead of evaluating sampled architectures. Our approach discovers the recurrent neural architecture within 0.78 GPU hour, which is 15.3x more efficient than ENAS [1] in terms of search time, and the architecture we discovered achieves the test perplexity of 56.1 on Penn Tree Bank (PTB) dataset, which is lower than ENAS by 2.2. In addition, we further demonstrate the usefulness of the learned architecture by transferring it to wiki-text-2 (WT2) dataset well. Moreover, the extended experiments on the WT2 dataset also show promising results.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Randomness-Enhanced Expressivity of Quantum Neural Networks
    Wu, Yadong
    Yao, Juan
    Zhang, Pengfei
    Li, Xiaopeng
    [J]. PHYSICAL REVIEW LETTERS, 2024, 132 (01)
  • [2] Training recurrent neural networks by using parallel tabu search algorithm based on crossover operation
    Kalinli, A
    Karaboga, D
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2004, 17 (05) : 529 - 542
  • [3] A tabu search algorithm for the training of neural networks
    Dengiz, B.
    Alabas-Uslu, C.
    Dengiz, O.
    [J]. JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2009, 60 (02) : 282 - 291
  • [4] A new architecture selection method based on tabu search for artificial neural networks
    Aladag, Cagdas Hakan
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (04) : 3287 - 3293
  • [5] Optimizing Deep Feedforward Neural Network Architecture: A Tabu Search Based Approach
    Tarun Kumar Gupta
    Khalid Raza
    [J]. Neural Processing Letters, 2020, 51 : 2855 - 2870
  • [6] Optimizing Deep Feedforward Neural Network Architecture: A Tabu Search Based Approach
    Gupta, Tarun Kumar
    Raza, Khalid
    [J]. NEURAL PROCESSING LETTERS, 2020, 51 (03) : 2855 - 2870
  • [7] Training recurrent neural networks for dynamic system identification using parallel tabu search algorithm
    Karaboga, D
    Kalinli, A
    [J]. PROCEEDINGS OF THE 1997 IEEE INTERNATIONAL SYMPOSIUM ON INTELLIGENT CONTROL, 1997, : 113 - 118
  • [8] Tabu Search Algorithm with Neural Tabu Mechanism for the Cyclic Job Shop Problem
    Bozejko, Wojciech
    Gnatowski, Andrzej
    Nizynski, Teodor
    Wodecki, Mieczyslaw
    [J]. ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING, (ICAISC 2016), PT II, 2016, 9693 : 409 - 418
  • [9] Neural architecture search algorithm based on voting scheme
    Yang, Jun
    Zhang, Jingfa
    [J]. Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2022, 30 (17): : 2119 - 2132
  • [10] Evolutionary Recurrent Neural Architecture Search
    Tian, Shuo
    Hu, Kai
    Guo, Shasha
    Li, Shiming
    Wang, Lei
    Xu, Weixia
    [J]. IEEE EMBEDDED SYSTEMS LETTERS, 2021, 13 (03) : 110 - 113