On a Sparse Shortcut Topology of Artificial Neural Networks

被引:8
|
作者
Fan F.-L. [1 ]
Wang D. [2 ]
Guo H. [1 ]
Zhu Q. [1 ]
Yan P. [1 ]
Wang G. [1 ]
Yu H. [2 ]
机构
[1] Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, 12180, NY
[2] University of Massachusetts, Department of Electrical and Computer Engineering, Lowell, 01854, MA
来源
关键词
Expressivity; generalizability; network architec- ture; shortcut network; theoretical deep learning;
D O I
10.1109/TAI.2021.3128132
中图分类号
学科分类号
摘要
In established network architectures, shortcut connections are often used to take the outputs of earlier layers as additional inputs to later layers. Despite the extraordinary effectiveness of shortcuts, there remain open questions on the mechanism and characteristics. For example, why are shortcuts powerful? Why do shortcuts generalize well? In this article, we investigate the expressivity and generalizability of a novel sparse shortcut topology. First, we demonstrate that this topology can empower a one-neuron-wide deep network to approximate any univariate continuous function. Then, we present a novel width-bounded universal approximator in contrast to depth-bounded universal approximators and extend the approximation result to a family of equally competent networks. Furthermore, with generalization bound theory, we show that the proposed shortcut topology enjoys excellent generalizability. Finally, we corroborate our theoretical analyses by comparing the proposed topology with popular architectures, including ResNet and DenseNet, on well-known benchmarks and perform a saliency map analysis to interpret the proposed topology. Our work helps understand the role of shortcuts and suggests further opportunities to innovate neural architectures. © 2020 IEEE.
引用
收藏
页码:595 / 608
页数:13
相关论文
共 50 条
  • [41] Topology of Deep Neural Networks
    Naitzat, Gregory
    Zhitnikov, Andrey
    Lim, Lek-Heng
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [42] Neural networks for topology optimization
    Sosnovik, Ivan
    Oseledets, Ivan
    RUSSIAN JOURNAL OF NUMERICAL ANALYSIS AND MATHEMATICAL MODELLING, 2019, 34 (04) : 215 - 223
  • [43] Optimal sparse network topology under sparse control in Laplacian networks
    Tang, Wentao
    Constantino, Pedro H.
    Daoutidis, Prodromos
    IFAC PAPERSONLINE, 2019, 52 (20): : 273 - 278
  • [44] Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science
    Decebal Constantin Mocanu
    Elena Mocanu
    Peter Stone
    Phuong H. Nguyen
    Madeleine Gibescu
    Antonio Liotta
    Nature Communications, 9
  • [45] Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science
    Mocanu, Decebal Constantin
    Mocanu, Elena
    Stone, Peter
    Nguyen, Phuong H.
    Gibescu, Madeleine
    Liotta, Antonio
    NATURE COMMUNICATIONS, 2018, 9
  • [46] Sparse sensor-based cylinder flow estimation using artificial neural networks
    Manohar, Kevin H.
    Morton, Chris
    Ziade, Paul
    PHYSICAL REVIEW FLUIDS, 2022, 7 (02)
  • [47] BRNeural - Artificial Neural Networks Simulator with Topology Multilayer Perceptron Using the Encog Framework
    Mathias, I. M.
    Junior, L. A. Z.
    Matyak, L. B.
    Dias, A. H.
    Duda, R. F.
    Afonso, G. M. S.
    IEEE LATIN AMERICA TRANSACTIONS, 2016, 14 (01) : 309 - 313
  • [48] Leak Detection and Topology Identification in Pipelines Using Fluid Transients and Artificial Neural Networks
    Bohorquez, Jessica
    Alexander, Bradley
    Simpson, Angus R.
    Lambert, Martin F.
    JOURNAL OF WATER RESOURCES PLANNING AND MANAGEMENT, 2020, 146 (06)
  • [49] Simulation based on artificial neural network for SBBR shortcut nitrification treatment
    College of Environmental Science and Engineering, Hunan University, Changsha 410082, China
    不详
    Zhongguo Huanjing Kexue/China Environmental Science, 2008, 28 (08): : 694 - 698
  • [50] Artificial neural networks
    Piuri, V
    Alippi, C
    JOURNAL OF SYSTEMS ARCHITECTURE, 1998, 44 (08) : 565 - 567