Empirical Loss Landscape Analysis of Neural Network Activation Functions

被引:1
|
作者
Bosman, Anna Sergeevna [1 ]
Engelbrecht, Andries [2 ,3 ]
Helbig, Marde [4 ]
机构
[1] Univ Pretoria, Dept Comp Sci, Pretoria, South Africa
[2] Univ Stellenbosch, Stellenbosch, South Africa
[3] Gulf Univ Sci & Technol, Ctr Appl Math & Bioinformat, Kuwait, Kuwait
[4] Griffith Univ, Sch Informat & Commun Technol, Southport, Qld, Australia
基金
新加坡国家研究基金会;
关键词
neural networks; activation functions; loss landscape; fitness landscape analysis;
D O I
10.1145/3583133.3596321
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Activation functions play a significant role in neural network design by enabling non-linearity. The choice of activation function was previously shown to influence the properties of the resulting loss landscape. Understanding the relationship between activation functions and loss landscape properties is important for neural architecture and training algorithm design. This study empirically investigates neural network loss landscapes associated with hyperbolic tangent, rectified linear unit, and exponential linear unit activation functions. Rectified linear unit is shown to yield the most convex loss landscape, and exponential linear unit is shown to yield the least flat loss landscape, and to exhibit superior generalisation performance. The presence of wide and narrow valleys in the loss landscape is established for all activation functions, and the narrow valleys are shown to correlate with saturated neurons and implicitly regularised network configurations.
引用
收藏
页码:2029 / 2037
页数:9
相关论文
共 50 条
  • [21] Recurrent neural network synthesis using interaction activation functions
    Novakovic, BM
    1996 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, PROCEEDINGS, VOLS 1-4, 1996, : 1608 - 1613
  • [22] Efficient Neural Network Robustness Certification with General Activation Functions
    Zhang, Huan
    Weng, Tsui-Wei
    Chen, Pin-Yu
    Hsieh, Cho-Jui
    Daniel, Luca
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [23] High precision FPGA implementation of neural network activation functions
    Ortega-Zamorano, Francisco
    Jerez, Jose M.
    Juarez, Gustavo
    Perez, Jorge O.
    Franco, Leonardo
    2014 IEEE SYMPOSIUM ON INTELLIGENT EMBEDDED SYSTEMS (IES), 2014, : 55 - 60
  • [24] EvoDNN - An Evolutionary Deep Neural Network with Heterogeneous Activation Functions
    Cui, Peiyu
    Shabash, Boris
    Wiese, Kay C.
    2019 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2019, : 2362 - 2369
  • [25] Study of the Effect of Combining Activation Functions in a Convolutional Neural Network
    Guevara, M.
    Cruz, V
    Vergara, O.
    Nandayapa, M.
    Ochoa, H.
    Sossa, H.
    IEEE LATIN AMERICA TRANSACTIONS, 2021, 19 (05) : 844 - 852
  • [26] Neuroevolutionary based convolutional neural network with adaptive activation functions
    ZahediNasab, Roxana
    Mohseni, Hadis
    NEUROCOMPUTING, 2020, 381 : 306 - 313
  • [27] The impact of activation functions on training and performance of a deep neural network
    Marcu, David C.
    Grava, Cristian
    2021 16TH INTERNATIONAL CONFERENCE ON ENGINEERING OF MODERN ELECTRIC SYSTEMS (EMES), 2021, : 126 - 129
  • [28] All-optical neural network with nonlinear activation functions
    Zuo, Ying
    Li, Bohan
    Zhao, Yujun
    Jiang, Yue
    Chen, You-Chiuan
    Chen, Peng
    Jo, Gyu-Boong
    Liu, Junwei
    Du, Shengwang
    OPTICA, 2019, 6 (09): : 1132 - 1137
  • [29] On Neural Network Activation Functions and Optimizers in Relation to Polynomial Regression
    Pomerat, John
    Segev, Aviv
    Datta, Rituparna
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 6183 - 6185
  • [30] New activation functions for single layer feedforward neural network
    Kocak, Yilmaz
    Siray, Gulesen Ustundag
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 164