Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions

被引:14
|
作者
Alkhouly, Asmaa A. [1 ]
Mohammed, Ammar [1 ,2 ]
Hefny, Hesham A. [1 ]
机构
[1] Cairo Univ, Fac Grad Studies Stat Res, Comp Sci Dept, Giza 12613, Egypt
[2] Misr Int Univ, Fac Comp Sci, Comp Sci Dept, Cairo 11865, Egypt
关键词
Neural networks; Neurons; Logistics; Convergence; Machine learning; Licenses; Computer science; Artificial neural network; deep neural network; learning challenges; activation function;
D O I
10.1109/ACCESS.2021.3085855
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.
引用
收藏
页码:82249 / 82271
页数:23
相关论文
共 50 条
  • [1] Improving the Performance of Neural Networks with an Ensemble of Activation Functions
    Nandi, Arijit
    Jana, Nanda Dulal
    Das, Swagatam
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [2] Improving the Accuracy of Deep Neural Networks Through Developing New Activation Functions
    Mercioni, Marina Adriana
    Tat, Angel Marcel
    Holban, Stefan
    [J]. 2020 IEEE 16TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTER COMMUNICATION AND PROCESSING (ICCP 2020), 2020, : 385 - 391
  • [3] Deep Neural Networks with Multistate Activation Functions
    Cai, Chenghao
    Xu, Yanyan
    Ke, Dengfeng
    Su, Kaile
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2015, 2015
  • [4] Activation Functions and Their Characteristics in Deep Neural Networks
    Ding, Bin
    Qian, Huimin
    Zhou, Jun
    [J]. PROCEEDINGS OF THE 30TH CHINESE CONTROL AND DECISION CONFERENCE (2018 CCDC), 2018, : 1836 - 1841
  • [5] A Formal Characterization of Activation Functions in Deep Neural Networks
    Amrouche, Massi
    Stipanovic, Dusan M.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 2153 - 2166
  • [6] Learning Activation Functions in Deep (Spline) Neural Networks
    Bohra, Pakshal
    Campos, Joaquim
    Gupta, Harshit
    Aziznejad, Shayan
    Unser, Michael
    [J]. IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2020, 1 : 295 - 309
  • [7] Adaptive Activation Functions for Skin Lesion Classification Using Deep Neural Networks
    Namozov, Abdulaziz
    Ergashev, Dilshod
    Cho, Young Im
    [J]. 2018 JOINT 10TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS (SCIS) AND 19TH INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS (ISIS), 2018, : 232 - 235
  • [8] Deep Kronecker neural networks: A general framework for neural networks with adaptive activation functions
    Jagtap, Ameya D.
    Shin, Yeonjong
    Kawaguchi, Kenji
    Karniadakis, George Em
    [J]. NEUROCOMPUTING, 2022, 468 : 165 - 180
  • [9] On the performance of pairings of activation and loss functions in neural networks
    Soares, Rodrigo G. F.
    Pereira, Enieson J. S.
    [J]. 2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 326 - 333
  • [10] Genetic Deep Neural Networks Using Different Activation Functions for Financial Data Mining
    Zhang, Luna M.
    [J]. PROCEEDINGS 2015 IEEE INTERNATIONAL CONFERENCE ON BIG DATA, 2015, : 2849 - 2851