Dropout Algorithms for Recurrent Neural Networks

被引:10
|
作者
Watt, Nathan [1 ]
du Plessis, Mathys C. [1 ]
机构
[1] Nelson Mandela Univ, Dept Comp Sci, POB 77000, ZA-6031 Port Elizabeth, South Africa
关键词
Deep Learning; Recurrent Neural Networks; Dropout;
D O I
10.1145/3278681.3278691
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the last decade, hardware advancements have allowed for neural networks to become much larger in size. Dropout is a popular deep learning technique which has shown to improve the performance of large neural networks. Recurrent neural networks are powerful networks specialised at solving problems which use time series data. Three different approaches to incorporating Dropout with recurrent neural networks have been suggested. However, these approaches have not been evaluated under identical experimental conditions. This article investigates the performance of these Dropout approaches using a 2D physics simulation benchmark. After applying statistical tests it was found that using Dropout did improve network performance on the benchmark. However, contrary to the literature, the Dropout approach which was expected to perform poorly, performed well, and the approach which was expected to perform well, performed poorly.
引用
下载
收藏
页码:72 / 78
页数:7
相关论文
共 50 条
  • [31] Variational Dropout Sparsifies Deep Neural Networks
    Molchanov, Dmitry
    Ashukha, Arsenii
    Vetrov, Dmitry
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [32] Dropout Rademacher complexity of deep neural networks
    Wei GAO
    Zhi-Hua ZHOU
    Science China(Information Sciences), 2016, 59 (07) : 173 - 184
  • [33] Regularization of deep neural networks with spectral dropout
    Khan, Salman H.
    Hayat, Munawar
    Porikli, Fatih
    NEURAL NETWORKS, 2019, 110 : 82 - 90
  • [34] Dropout Rademacher complexity of deep neural networks
    Wei Gao
    Zhi-Hua Zhou
    Science China Information Sciences, 2016, 59
  • [35] Analysis on the Dropout Effect in Convolutional Neural Networks
    Park, Sungheon
    Kwak, Nojun
    COMPUTER VISION - ACCV 2016, PT II, 2017, 10112 : 189 - 204
  • [36] A General Approach to Dropout in Quantum Neural Networks
    Scala, Francesco
    Ceschini, Andrea
    Panella, Massimo
    Gerace, Dario
    ADVANCED QUANTUM TECHNOLOGIES, 2023,
  • [37] Dropout Rademacher complexity of deep neural networks
    Gao, Wei
    Zhou, Zhi-Hua
    SCIENCE CHINA-INFORMATION SCIENCES, 2016, 59 (07)
  • [38] Two constructive algorithms for improved time series processing with recurrent neural networks
    Bone, Romuald
    Crucianu, Michel
    de Beauville, Jean-Pierre Asselin
    Neural Networks for Signal Processing - Proceedings of the IEEE Workshop, 2000, 1 : 55 - 64
  • [39] Constrained formulations and algorithms for predicting stock prices by recurrent fir neural networks
    Wah, Benjamin W.
    Qian, Ming-Lun
    INTERNATIONAL JOURNAL OF INFORMATION TECHNOLOGY & DECISION MAKING, 2006, 5 (04) : 639 - 658
  • [40] PARAMETER GENERATION ALGORITHMS FOR TEXT-TO-SPEECH SYNTHESIS WITH RECURRENT NEURAL NETWORKS
    Klimkov, Viacheslav
    Moinet, Alexis
    Nadolski, Adam
    Drugman, Thomas
    2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 626 - 631