On a Sparse Shortcut Topology of Artificial Neural Networks

被引:8
|
作者
Fan F.-L. [1 ]
Wang D. [2 ]
Guo H. [1 ]
Zhu Q. [1 ]
Yan P. [1 ]
Wang G. [1 ]
Yu H. [2 ]
机构
[1] Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, 12180, NY
[2] University of Massachusetts, Department of Electrical and Computer Engineering, Lowell, 01854, MA
来源
关键词
Expressivity; generalizability; network architec- ture; shortcut network; theoretical deep learning;
D O I
10.1109/TAI.2021.3128132
中图分类号
学科分类号
摘要
In established network architectures, shortcut connections are often used to take the outputs of earlier layers as additional inputs to later layers. Despite the extraordinary effectiveness of shortcuts, there remain open questions on the mechanism and characteristics. For example, why are shortcuts powerful? Why do shortcuts generalize well? In this article, we investigate the expressivity and generalizability of a novel sparse shortcut topology. First, we demonstrate that this topology can empower a one-neuron-wide deep network to approximate any univariate continuous function. Then, we present a novel width-bounded universal approximator in contrast to depth-bounded universal approximators and extend the approximation result to a family of equally competent networks. Furthermore, with generalization bound theory, we show that the proposed shortcut topology enjoys excellent generalizability. Finally, we corroborate our theoretical analyses by comparing the proposed topology with popular architectures, including ResNet and DenseNet, on well-known benchmarks and perform a saliency map analysis to interpret the proposed topology. Our work helps understand the role of shortcuts and suggests further opportunities to innovate neural architectures. © 2020 IEEE.
引用
收藏
页码:595 / 608
页数:13
相关论文
共 50 条
  • [31] Topology and weight evolving artificial neural networks in cooperative transport by a robotic swarm
    Motoaki Hiraga
    Kazuhiro Ohkura
    Artificial Life and Robotics, 2022, 27 : 324 - 332
  • [32] Compact yet efficient hardware implementation of artificial neural networks with customized topology
    Nedjah, Nadia
    da Silva, Rodrigo Martins
    Mourelle, Luiza de Macedo
    EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (10) : 9191 - 9206
  • [33] Robust topology optimization with low rank approximation using artificial neural networks
    Keshavarzzadeh, Vahid
    Kirby, Robert M.
    Narayan, Akil
    COMPUTATIONAL MECHANICS, 2021, 68 (06) : 1297 - 1323
  • [34] Combining theoretical knowledge and artificial neural networks for power system topology verification
    Lukomski, R
    Wilkosz, K
    ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 2294 - 2298
  • [35] The equality of the Patch topology and the Ultrafilter topology: A shortcut
    Ruza, Luz M.
    Vielma, Jorge
    APPLIED GENERAL TOPOLOGY, 2011, 12 (01): : 15 - 16
  • [36] SPARSE TOPOLOGY IDENTIFICATION FOR POINT PROCESS NETWORKS
    Pasha, Syed Ahmed
    Solo, Victor
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2196 - 2200
  • [37] Classification by Sparse Neural Networks
    Kurkova, Vera
    Sanguineti, Marcello
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (09) : 2746 - 2754
  • [38] Training Sparse Neural Networks
    Srinivas, Suraj
    Subramanya, Akshayvarun
    Babu, R. Venkatesh
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 455 - 462
  • [39] Sparse Convolutional Neural Networks
    Lu, Haoyuan
    Wang, Min
    Foroosh, Hassan
    Tappen, Marshall
    Penksy, Marianna
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 806 - 814
  • [40] Topology of deep neural networks
    1600, Microtome Publishing (21):