Empirical Loss Landscape Analysis of Neural Network Activation Functions

被引:1
|
作者
Bosman, Anna Sergeevna [1 ]
Engelbrecht, Andries [2 ,3 ]
Helbig, Marde [4 ]
机构
[1] Univ Pretoria, Dept Comp Sci, Pretoria, South Africa
[2] Univ Stellenbosch, Stellenbosch, South Africa
[3] Gulf Univ Sci & Technol, Ctr Appl Math & Bioinformat, Kuwait, Kuwait
[4] Griffith Univ, Sch Informat & Commun Technol, Southport, Qld, Australia
基金
新加坡国家研究基金会;
关键词
neural networks; activation functions; loss landscape; fitness landscape analysis;
D O I
10.1145/3583133.3596321
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Activation functions play a significant role in neural network design by enabling non-linearity. The choice of activation function was previously shown to influence the properties of the resulting loss landscape. Understanding the relationship between activation functions and loss landscape properties is important for neural architecture and training algorithm design. This study empirically investigates neural network loss landscapes associated with hyperbolic tangent, rectified linear unit, and exponential linear unit activation functions. Rectified linear unit is shown to yield the most convex loss landscape, and exponential linear unit is shown to yield the least flat loss landscape, and to exhibit superior generalisation performance. The presence of wide and narrow valleys in the loss landscape is established for all activation functions, and the narrow valleys are shown to correlate with saturated neurons and implicitly regularised network configurations.
引用
收藏
页码:2029 / 2037
页数:9
相关论文
共 50 条
  • [1] Empirical Evaluation of Activation Functions in Deep Convolution Neural Network for Facial Expression Recognition
    Khalid, Muhammad
    Baber, Junaid
    Kasi, Mumraiz Khan
    Bakhtyar, Maheen
    Devi, Varsha
    Sheikh, Naveed
    2020 43RD INTERNATIONAL CONFERENCE ON TELECOMMUNICATIONS AND SIGNAL PROCESSING (TSP), 2020, : 204 - 207
  • [2] The Role of Neural Network Activation Functions
    Parhi, Rahul
    Nowak, Robert D.
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 (1779-1783) : 1779 - 1783
  • [3] Learning algorithm analysis for deep neural network with ReLu activation functions
    Placzek, Stanislaw
    Placzek, Aleksander
    COMPUTER APPLICATIONS IN ELECTRICAL ENGINEERING (ZKWE'2018), 2018, 19
  • [4] On the performance of pairings of activation and loss functions in neural networks
    Soares, Rodrigo G. F.
    Pereira, Enieson J. S.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 326 - 333
  • [5] The loss surfaces of neural networks with general activation functions
    Baskerville, Nicholas P.
    Keating, Jonathan P.
    Mezzadri, Francesco
    Najnudel, Joseph
    JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2021, 2021 (06):
  • [6] Empirical loss landscape analysis in deep learning: A survey
    Liang R.
    Liu B.
    Sun Y.
    Xitong Gongcheng Lilun yu Shijian/System Engineering Theory and Practice, 2023, 43 (03): : 813 - 823
  • [7] Is the Skip Connection Provable to Reform the Neural Network Loss Landscape?
    Wang, Lifu
    Shen, Bo
    Zhao, Ning
    Zhang, Zhiyuan
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2792 - 2798
  • [8] Neural network with unbounded activation functions is universal approximator
    Sonoda, Sho
    Murata, Noboru
    APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2017, 43 (02) : 233 - 268
  • [9] Resource efficient activation functions for neural network accelerators
    Wuraola, Adedamola
    Patel, Nitish
    NEUROCOMPUTING, 2022, 482 : 163 - 185
  • [10] Fusion of Chaotic Activation Functions in Training Neural Network
    Kabir, A. N. M. Enamul
    Uddin, A. F. M. Nokib
    Asaduzzaman, M.
    Hasan, M. Fahad
    Hasan, M. Imran
    Shahjahan, Md
    2012 7TH INTERNATIONAL CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (ICECE), 2012,