Learning deep hierarchical and temporal recurrent neural networks with residual learning

被引:1
|
作者
Tehseen Zia
Assad Abbas
Usman Habib
Muhammad Sajid Khan
机构
[1] COMSATS University Islamabad,Department of Computer Science
[2] National University of Computer and Emerging Sciences (FAST-NUCES),College of Computer Science
[3] Sichuan University,undefined
关键词
Deep learning; Recurrent neural networks; Residual learning; Long-short term memory; Sequence modeling;
D O I
暂无
中图分类号
学科分类号
摘要
Learning both hierarchical and temporal dependencies can be crucial for recurrent neural networks (RNNs) to deeply understand sequences. To this end, a unified RNN framework is required that can ease the learning of both the deep hierarchical and temporal structures by allowing gradients to propagate back from both ends without being vanished. The residual learning (RL) has appeared as an effective and less-costly method to facilitate backward propagation of gradients. The significance of the RL is exclusively shown for learning deep hierarchical representations and temporal dependencies. Nevertheless, there is lack of efforts to unify these finding into a single framework for learning deep RNNs. In this study, we aim to prove that approximating identity mapping is crucial for optimizing both hierarchical and temporal structures. We propose a framework called hierarchical and temporal residual RNNs, to learn RNNs by approximating identity mappings across hierarchical and temporal structures. To validate the proposed method, we explore the efficacy of employing shortcut connections for training deep RNNs structures for sequence learning problems. Experiments are performed on Penn Treebank, Hutter Prize and IAM-OnDB datasets and results demonstrate the utility of the framework in terms of accuracy and computational complexity. We demonstrate that even for large datasets exploiting parameters for increasing network depth can gain computational benefits with reduced size of the RNN "state".
引用
收藏
页码:873 / 882
页数:9
相关论文
共 50 条
  • [31] Learning Queuing Networks by Recurrent Neural Networks
    Garbi, Giulio
    Incerto, Emilio
    Tribastone, Mirco
    PROCEEDINGS OF THE ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING (ICPE'20), 2020, : 56 - 66
  • [32] Online Deep Learning: Learning Deep Neural Networks on the Fly
    Sahoo, Doyen
    Pham, Quang
    Lu, Jing
    Hoi, Steven C. H.
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2660 - 2666
  • [33] Learning with recurrent neural networks - Conclusion
    不详
    LEARNING WITH RECURRENT NEURAL NETWORKS, 2000, 254 : 133 - 135
  • [34] Learning with recurrent neural networks - Introduction
    Hammer, B
    LEARNING WITH RECURRENT NEURAL NETWORKS, 2000, 254 : 1 - +
  • [35] LEARNING COMPACT RECURRENT NEURAL NETWORKS
    Lu, Zhiyun
    Sindhwani, Vikas
    Sainath, Tara N.
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 5960 - 5964
  • [36] Bayesian learning for recurrent neural networks
    Crucianu, M
    Boné, R
    de Beauville, JPA
    NEUROCOMPUTING, 2001, 36 (01) : 235 - 242
  • [37] Hierarchical Meta-learning Models with Deep Neural Networks for Spectrum Assignment
    Rutagemwa, Humphrey
    Baddour, Kareem E.
    Rong, Bo
    2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2019,
  • [38] Hierarchical Meta-learning Models with Deep Neural Networks for Spectrum Assignment
    Rutagemwa, Humphrey
    Baddour, Kareem E.
    Rong, Bo
    2019 IEEE PACIFIC RIM CONFERENCE ON COMMUNICATIONS, COMPUTERS AND SIGNAL PROCESSING (PACRIM), 2019,
  • [39] SlumberNet: deep learning classification of sleep stages using residual neural networks
    Pawan K. Jha
    Utham K. Valekunja
    Akhilesh B. Reddy
    Scientific Reports, 14
  • [40] An improved model training method for residual convolutional neural networks in deep learning
    Li, Xuelei
    Li, Rengang
    Zhao, Yaqian
    Zhao, Jian
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (05) : 6811 - 6821