Theory of deep convolutional neural networks: Downsampling

被引:134
|
作者
Zhou, Ding-Xuan [1 ,2 ]
机构
[1] City Univ Hong Kong, Sch Data Sci, Kowloon, Hong Kong, Peoples R China
[2] City Univ Hong Kong, Dept Math, Kowloon, Hong Kong, Peoples R China
关键词
Deep learning; Convolutional neural networks; Approximation theory; Downsampling; Filter masks; MULTILAYER FEEDFORWARD NETWORKS; OPTIMAL APPROXIMATION; REGRESSION; ALGORITHM; BOUNDS;
D O I
10.1016/j.neunet.2020.01.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Establishing a solid theoretical foundation for structured deep neural networks is greatly desired due to the successful applications of deep learning in various practical domains. This paper aims at an approximation theory of deep convolutional neural networks whose structures are induced by convolutions. To overcome the difficulty in theoretical analysis of the networks with linearly increasing widths arising from convolutions, we introduce a downsampling operator to reduce the widths. We prove that the downsampled deep convolutional neural networks can be used to approximate ridge functions nicely, which hints some advantages of these structured networks in terms of approximation or modeling. We also prove that the output of any multi-layer fully-connected neural network can be realized by that of a downsampled deep convolutional neural network with free parameters of the same order, which shows that in general, the approximation ability of deep convolutional neural networks is at least as good as that of fully-connected networks. Finally, a theorem for approximating functions on Riemannian manifolds is presented, which demonstrates that deep convolutional neural networks can be used to learn manifold features of data. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页码:319 / 327
页数:9
相关论文
共 50 条
  • [41] IMPROVEMENTS TO DEEP CONVOLUTIONAL NEURAL NETWORKS FOR LVCSR
    Sainath, Tara N.
    Kingsbury, Brian
    Mohamed, Abdel-rahman
    Dahl, George E.
    Saon, George
    Soltau, Hagen
    Beran, Tomas
    Aravkin, Aleksandr Y.
    Ramabhadran, Bhuvana
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 315 - 320
  • [42] Deep Convolutional Neural Networks for pedestrian detection
    Tome, D.
    Monti, F.
    Baroffio, L.
    Bondi, L.
    Tagliasacchi, M.
    Tubaro, S.
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2016, 47 : 482 - 489
  • [43] Stenosis Detection with Deep Convolutional Neural Networks
    Antczak, Karol
    Liberadzki, Lukasz
    22ND INTERNATIONAL CONFERENCE ON CIRCUITS, SYSTEMS, COMMUNICATIONS AND COMPUTERS (CSCC 2018), 2018, 210
  • [44] A Novel Connectivity of Deep Convolutional Neural Networks
    Shen, Zhixi
    Liu, Yong
    2017 CHINESE AUTOMATION CONGRESS (CAC), 2017, : 7779 - 7783
  • [45] xDNN: Inference for Deep Convolutional Neural Networks
    D'Alberto, Paolo
    Wu, Victor
    Ng, Aaron
    Nimaiyar, Rahul
    Delaye, Elliott
    Sirasao, Ashish
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2022, 15 (02)
  • [46] Survey on Deep Convolutional Neural Networks in Mammography
    Abdelhafiz, Dina
    Nabavi, Sheida
    Ammar, Reda
    Yang, Clifford
    2017 IEEE 7TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL ADVANCES IN BIO AND MEDICAL SCIENCES (ICCABS), 2017,
  • [47] Information Bottleneck Theory on Convolutional Neural Networks
    Junjie Li
    Ding Liu
    Neural Processing Letters, 2021, 53 : 1385 - 1400
  • [48] Deep convolutional neural networks for data delivery in vehicular networks
    Jiang, Hejun
    Tang, Xiaolan
    Jin, Kai
    Chen, Wenlong
    Pu, Juhua
    NEUROCOMPUTING, 2021, 432 (432) : 216 - 226
  • [49] Information Bottleneck Theory on Convolutional Neural Networks
    Li, Junjie
    Liu, Ding
    NEURAL PROCESSING LETTERS, 2021, 53 (02) : 1385 - 1400
  • [50] Convolutional modulation theory: A bridge between convolutional neural networks and signal modulation theory
    Wu, Fuzhi
    Wu, Jiasong
    Kong, Youyong
    Yang, Chunfeng
    Yang, Guanyu
    Shu, Huazhong
    Carrault, Guy
    Senhadji, Lotfi
    NEUROCOMPUTING, 2022, 514 : 195 - 215