Understanding Weight Normalized Deep Neural Networks with Rectified Linear Units

被引:0
|
作者
Xu, Yixi [1 ]
Wang, Xiao [1 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47907 USA
关键词
CLASSIFICATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a general framework for norm-based capacity control for L-p,(q) weight normalized deep neural networks. We establish the upper bound on the Rademacher complexities of this family. With an L-p,(q) normalization where q <= p* and 1/p +1/p* = 1, we discuss properties of a width-independent capacity control, which only depends on the depth by a square root term. We further analyze the approximation properties of L-p,(q) weight normalized deep neural networks. In particular, for an L-i,L-infinity weight normalized network, the approximation error can be controlled by the L-1 norm of the output layer, and the corresponding generalization error only depends on the architecture by the square root of the depth.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Deep Learning with S-Shaped Rectified Linear Activation Units
    Jin, Xiaojie
    Xu, Chunyan
    Feng, Jiashi
    Wei, Yunchao
    Xiong, Junjun
    Yan, Shuicheng
    [J]. THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1737 - 1743
  • [22] Dual Rectified Linear Units (DReLUs): A replacement for tanh activation functions in Quasi-Recurrent Neural Networks
    Godin, Frederic
    Degrave, Jonas
    Dambre, Joni
    De Neye, Wesley
    [J]. PATTERN RECOGNITION LETTERS, 2018, 116 : 8 - 14
  • [23] A Dynamic Rectified Linear Activation Units
    Hu, Xiaobin
    Niu, Peifeng
    Wang, Jianmei
    Zhang, Xinxin
    [J]. IEEE ACCESS, 2019, 7 : 180409 - 180416
  • [24] Image Denoising with Rectified Linear Units
    Wu, Yangwei
    Zhao, Haohua
    Zhang, Liqing
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2014, PT III, 2014, 8836 : 142 - 149
  • [25] ON RECTIFIED LINEAR UNITS FOR SPEECH PROCESSING
    Zeiler, M. D.
    Ranzato, M.
    Monga, R.
    Mao, M.
    Yang, K.
    Le, Q. V.
    Nguyen, P.
    Senior, A.
    Vanhoucke, V.
    Dean, J.
    Hinton, G. E.
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 3517 - 3521
  • [26] The Vietnamese Speech Recognition Based on Rectified Linear Units Deep Neural Network and Spoken Term Detection System Combination
    Xiong, Shifu
    Guo, Wu
    Liu, Diyuan
    [J]. 2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 183 - 186
  • [27] Rectified Linear Neural Networks with Tied-Scalar Regularization for LVCSR
    Zhang, Shiliang
    Jiang, Hui
    Wei, Si
    Dai, Li-Rong
    [J]. 16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2635 - 2639
  • [28] Rectified-linear and Recurrent Neural Networks Built with Spin Devices
    Dong, Qing
    Yang, Kaiyuan
    Fick, Laura
    Blaauw, David
    Sylvester, Dennis
    [J]. 2017 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2017, : 2492 - 2495
  • [29] Graph-adaptive Rectified Linear Unit for Graph Neural Networks
    Zhang, Yifei
    Zhu, Hao
    Meng, Ziqiao
    Koniusz, Piotr
    King, Irwin
    [J]. PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 1331 - 1339
  • [30] Memory Capacity of Neural Networks with Threshold and Rectified Linear Unit Activations
    Vershynin, Roman
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (04): : 1004 - 1033