Residual Networks of Residual Networks: Multilevel Residual Networks

被引:205
|
作者
Zhang, Ke [1 ]
Sun, Miao [2 ]
Han, Tony X. [2 ]
Yuan, Xingfang [2 ]
Guo, Liru [1 ]
Liu, Tao [1 ]
机构
[1] North China Elect Power Univ, Dept Elect & Commun Engn, Baoding 071000, Hebei, Peoples R China
[2] Univ Missouri, Dept Elect & Comp Engn, Columiba, MO 65211 USA
基金
中国国家自然科学基金;
关键词
Image classification; ImageNet data set; residual networks; residual networks of residual networks (RoR); shortcut; stochastic depth (SD);
D O I
10.1109/TCSVT.2017.2654543
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
A residual networks family with hundreds or even thousands of layers dominates major image recognition tasks, but building a network by simply stacking residual blocks inevitably limits its optimization ability. This paper proposes a novel residual network architecture, residual networks of residual networks (RoR), to dig the optimization ability of residual networks. RoR substitutes optimizing residual mapping of residual mapping for optimizing original residual mapping. In particular, RoR adds levelwise shortcut connections upon original residual networks to promote the learning capability of residual networks. More importantly, RoR can be applied to various kinds of residual networks (ResNets, Pre-ResNets, and WRN) and significantly boost their performance. Our experiments demonstrate the effectiveness and versatility of RoR, where it achieves the best performance in all residual-networklike structures. Our RoR-3-WRN58-4 + SD models achieve new state-of-the-art results on CIFAR-10, CIFAR-100, and SVHN, with the test errors of 3.77%, 19.73%, and 1.59%, respectively. RoR-3 models also achieve state-of-the-art results compared with ResNets on the ImageNet data set.
引用
收藏
页码:1303 / 1314
页数:12
相关论文
共 50 条
  • [41] Parameters Sharing in Residual Neural Networks
    Dai, Dawei
    Yu, Liping
    Wei, Hui
    [J]. NEURAL PROCESSING LETTERS, 2020, 51 (02) : 1393 - 1410
  • [42] LDS-Inspired Residual Networks
    Dimou, Anastasios
    Ataloglou, Dimitrios
    Dimitropoulos, Kosmas
    Alvarez, Federico
    Daras, Petros
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (08) : 2363 - 2375
  • [43] Residual Neural Networks for Digital Predistortion
    Wu, Yibo
    Gustavsson, Ulf
    Amat, Alexandre Graell, I
    Wymeersch, Henk
    [J]. 2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [44] Residual Closeness in Cycles and Related Networks
    Odabas, Zeynep Nihan
    Aytac, Aysun
    [J]. FUNDAMENTA INFORMATICAE, 2013, 124 (03) : 297 - 307
  • [45] Parameters Sharing in Residual Neural Networks
    Dawei Dai
    Liping Yu
    Hui Wei
    [J]. Neural Processing Letters, 2020, 51 : 1393 - 1410
  • [46] Deep limits of residual neural networks
    Thorpe, Matthew
    van Gennip, Yves
    [J]. RESEARCH IN THE MATHEMATICAL SCIENCES, 2023, 10 (01)
  • [47] Scaling Properties of Deep Residual Networks
    Cohen, Alain-Sam
    Cont, Rama
    Rossier, Alain
    Xu, Renyuan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [48] Identity Mappings in Deep Residual Networks
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 : 630 - 645
  • [49] Study of Residual Networks for Image Recognition
    Ebrahimi, Mohammad Sadegh
    Abadi, Hossein Karkeh
    [J]. INTELLIGENT COMPUTING, VOL 2, 2021, 284 : 754 - 763
  • [50] Residual Networks Behave Like Ensembles of Relatively Shallow Networks
    Veit, Andreas
    Wilber, Michael
    Belongie, Serge
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29