NORMALIZATION EFFECTS ON DEEP NEURAL NETWORKS

被引:1
|
作者
Yu, Jiahui [1 ]
Spiliopoulos, Konstantinos [1 ]
机构
[1] Boston Univ, Dept Math & Stat, Boston, MA 02215 USA
来源
FOUNDATIONS OF DATA SCIENCE | 2023年 / 5卷 / 03期
基金
美国国家科学基金会;
关键词
Machine learning; neural networks; normalization effect; scaling effects; asymptotic expansions; out-of-sample performance; APPROXIMATION;
D O I
10.3934/fods.2023004
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We study the effect of normalization on the layers of deep neural networks of feed-forward type. A given layer i with Ni hidden units is allowed to be normalized by 1/N gamma i i with-yi is an element of [1/2, 1] and we study the effect of the choice of the-yi on the statistical behavior of the neural network's output (such as variance) as well as on the test accuracy on the MNIST data set. We find that in terms of variance of the neural network's output and test accuracy the best choice is to choose the-yi's to be equal to one, which is the mean-field scaling. We also find that this is particularly true for the outer layer, in that the neural network's behavior is more sensitive in the scaling of the outer layer as opposed to the scaling of the inner layers. The mechanism for the mathematical analysis is an asymptotic expansion for the neural network's output. An important practical consequence of the analysis is that it provides a systematic and mathematically informed way to choose the learning rate hyperparameters. Such a choice guarantees that the neural network behaves in a statistically robust way as the Ni grow to infinity.
引用
收藏
页码:389 / 465
页数:77
相关论文
共 50 条
  • [1] Is normalization indispensable for training deep neural networks?
    Shao, Jie
    Hu, Kai
    Wang, Changhu
    Xue, Xiangyang
    Raj, Bhiksha
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [2] On Centralization and Unitization of Batch Normalization for Deep ReLU Neural Networks
    Fei, Wen
    Dai, Wenrui
    Li, Chenglin
    Zou, Junni
    Xiong, Hongkai
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 2827 - 2841
  • [3] Accelerating Spectral Normalization for Enhancing Robustness of Deep Neural Networks
    Pan, Zhixin
    Mishra, Prabhat
    [J]. 2021 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2021), 2021, : 260 - 265
  • [4] Scaling-Based Weight Normalization for Deep Neural Networks
    Yuan, Qunyong
    Xiao, Nanfeng
    [J]. IEEE ACCESS, 2019, 7 : 7286 - 7295
  • [5] Generalized Batch Normalization: Towards Accelerating Deep Neural Networks
    Yuan, Xiaoyong
    Feng, Zheng
    Norton, Matthew
    Li, Xiaolin
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 1682 - 1689
  • [6] Improving Batch Normalization with Skewness Reduction for Deep Neural Networks
    Ding, Pak Lun Kevin
    Martin, Sarah
    Li, Baoxin
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7165 - 7172
  • [7] Transferable Normalization: Towards Improving Transferability of Deep Neural Networks
    Wang, Ximei
    Jin, Ying
    Long, Mingsheng
    Wang, Jianmin
    Jordan, Michael I.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [8] Mean Spectral Normalization of Deep Neural Networks for Embedded Automation
    Subramanian, Anand Krishnamoorthy
    Chong, Nak Young
    [J]. 2019 IEEE 15TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2019, : 249 - 256
  • [9] Centered Weight Normalization in Accelerating Training of Deep Neural Networks
    Huang, Lei
    Liu, Xianglong
    Liu, Yang
    Lang, Bo
    Tao, Dacheng
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2822 - 2830
  • [10] Normalization of photoplethysmography using deep neural networks for individual and group comparison
    Kim, Ji Woon
    Choi, Seong-Wook
    [J]. INTERFACE-COMUNICACAO SAUDE EDUCACAO, 2022, 26