Dropout Algorithms for Recurrent Neural Networks

被引:10
|
作者
Watt, Nathan [1 ]
du Plessis, Mathys C. [1 ]
机构
[1] Nelson Mandela Univ, Dept Comp Sci, POB 77000, ZA-6031 Port Elizabeth, South Africa
关键词
Deep Learning; Recurrent Neural Networks; Dropout;
D O I
10.1145/3278681.3278691
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the last decade, hardware advancements have allowed for neural networks to become much larger in size. Dropout is a popular deep learning technique which has shown to improve the performance of large neural networks. Recurrent neural networks are powerful networks specialised at solving problems which use time series data. Three different approaches to incorporating Dropout with recurrent neural networks have been suggested. However, these approaches have not been evaluated under identical experimental conditions. This article investigates the performance of these Dropout approaches using a 2D physics simulation benchmark. After applying statistical tests it was found that using Dropout did improve network performance on the benchmark. However, contrary to the literature, the Dropout approach which was expected to perform poorly, performed well, and the approach which was expected to perform well, performed poorly.
引用
下载
收藏
页码:72 / 78
页数:7
相关论文
共 50 条
  • [41] Recurrent neural networks for reinforcement learning: Architecture, learning algorithms and internal representation
    Onat, A
    Kita, H
    Nishikawa, Y
    IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, 1998, : 2010 - 2015
  • [42] Kurtosis-Based CRTRL Algorithms for Fully Connected Recurrent Neural Networks
    Menguc, Engin Cemal
    Acir, Nurettin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (12) : 6123 - 6131
  • [43] Two constructive algorithms for improved time series processing with recurrent neural networks
    Boné, R
    Crucianu, M
    de Beauville, JPA
    NEURAL NETWORKS FOR SIGNAL PROCESSING X, VOLS 1 AND 2, PROCEEDINGS, 2000, : 55 - 64
  • [44] Recurrent neural networks
    Siegelmann, HT
    COMPUTER SCIENCE TODAY, 1995, 1000 : 29 - 45
  • [45] Recognizing recurrent neural networks (rRNN): Bayesian inference for recurrent neural networks
    Sebastian Bitzer
    Stefan J. Kiebel
    Biological Cybernetics, 2012, 106 : 201 - 217
  • [46] Recognizing recurrent neural networks (rRNN): Bayesian inference for recurrent neural networks
    Bitzer, Sebastian
    Kiebel, Stefan J.
    BIOLOGICAL CYBERNETICS, 2012, 106 (4-5) : 201 - 217
  • [47] Revisiting spatial dropout for regularizing convolutional neural networks
    Lee, Sanghun
    Lee, Chulhee
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (45-46) : 34195 - 34207
  • [48] Revisiting spatial dropout for regularizing convolutional neural networks
    Sanghun Lee
    Chulhee Lee
    Multimedia Tools and Applications, 2020, 79 : 34195 - 34207
  • [49] New second-order algorithms for recurrent neural networks based on conjugate gradient
    Campolucci, P
    Simonetti, M
    Uncini, A
    Piazza, F
    IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, 1998, : 384 - 389
  • [50] Unemployment Rate Prediction Using a Hybrid Model of Recurrent Neural Networks and Genetic Algorithms
    Mero, Kevin
    Salgado, Nelson
    Meza, Jaime
    Pacheco-Delgado, Janeth
    Ventura, Sebastian
    APPLIED SCIENCES-BASEL, 2024, 14 (08):