On a Sparse Shortcut Topology of Artificial Neural Networks

被引:8
|
作者
Fan F.-L. [1 ]
Wang D. [2 ]
Guo H. [1 ]
Zhu Q. [1 ]
Yan P. [1 ]
Wang G. [1 ]
Yu H. [2 ]
机构
[1] Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, 12180, NY
[2] University of Massachusetts, Department of Electrical and Computer Engineering, Lowell, 01854, MA
来源
关键词
Expressivity; generalizability; network architec- ture; shortcut network; theoretical deep learning;
D O I
10.1109/TAI.2021.3128132
中图分类号
学科分类号
摘要
In established network architectures, shortcut connections are often used to take the outputs of earlier layers as additional inputs to later layers. Despite the extraordinary effectiveness of shortcuts, there remain open questions on the mechanism and characteristics. For example, why are shortcuts powerful? Why do shortcuts generalize well? In this article, we investigate the expressivity and generalizability of a novel sparse shortcut topology. First, we demonstrate that this topology can empower a one-neuron-wide deep network to approximate any univariate continuous function. Then, we present a novel width-bounded universal approximator in contrast to depth-bounded universal approximators and extend the approximation result to a family of equally competent networks. Furthermore, with generalization bound theory, we show that the proposed shortcut topology enjoys excellent generalizability. Finally, we corroborate our theoretical analyses by comparing the proposed topology with popular architectures, including ResNet and DenseNet, on well-known benchmarks and perform a saliency map analysis to interpret the proposed topology. Our work helps understand the role of shortcuts and suggests further opportunities to innovate neural architectures. © 2020 IEEE.
引用
收藏
页码:595 / 608
页数:13
相关论文
共 50 条
  • [1] Sparse solution in training artificial neural networks
    Giustolisi, O
    NEUROCOMPUTING, 2004, 56 : 285 - 304
  • [2] On the use of artificial neural networks in topology optimisation
    Woldseth, Rebekka V.
    Aage, Niels
    Baerentzen, J. Andreas
    Sigmund, Ole
    STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION, 2022, 65 (10)
  • [3] ARTIFICIAL NEURAL NETWORKS IN THE ANALYSIS OF BEHAVIORAL TOPOLOGY
    KORZ, V
    SCHADE, U
    LAUBENSTEIN, U
    HENDRICHS, H
    NATURWISSENSCHAFTEN, 1995, 82 (10) : 479 - 481
  • [4] Learnt Topology Gating Artificial Neural Networks
    Kadlec, Petr
    Gabrys, Bogdan
    2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2604 - 2611
  • [5] On the use of artificial neural networks in topology optimisation
    Rebekka V. Woldseth
    Niels Aage
    J. Andreas Bærentzen
    Ole Sigmund
    Structural and Multidisciplinary Optimization, 2022, 65
  • [6] Influence of random topology in artificial neural networks: A survey
    Kaviani, Sara
    Sohn, Insoo
    ICT EXPRESS, 2020, 6 (02): : 145 - 150
  • [7] Artificial neural networks in power system topology recognition
    Delimar, M
    Pavic, I
    Hebel, Z
    IEEE REGION 8 EUROCON 2003, VOL B, PROCEEDINGS: COMPUTER AS A TOOL, 2003, : 287 - 291
  • [8] Shortcut learning in deep neural networks
    Geirhos, Robert
    Jacobsen, Joern-Henrik
    Michaelis, Claudio
    Zemel, Richard
    Brendel, Wieland
    Bethge, Matthias
    Wichmann, Felix A.
    NATURE MACHINE INTELLIGENCE, 2020, 2 (11) : 665 - 673
  • [9] Shortcut learning in deep neural networks
    Robert Geirhos
    Jörn-Henrik Jacobsen
    Claudio Michaelis
    Richard Zemel
    Wieland Brendel
    Matthias Bethge
    Felix A. Wichmann
    Nature Machine Intelligence, 2020, 2 : 665 - 673
  • [10] Efficient Bayesian Learning of Sparse Deep Artificial Neural Networks
    Fakhfakh, Mohamed
    Bouaziz, Bassem
    Chaari, Lotfi
    Gargouri, Faiez
    ADVANCES IN INTELLIGENT DATA ANALYSIS XX, IDA 2022, 2022, 13205 : 78 - 88