Review of Adaptive Activation Function in Deep Neural Network

被引:0
|
作者
Lau, Mian Mian [1 ]
Lim, King Hann [1 ]
机构
[1] Curtin Univ Malaysia, Curtin Sarawak Res Inst, CDT 250, Miri Sarawak, Malaysia
关键词
Adaptive activation functions; Saturated activation fucntions; Unsaturated actuvation fucntions; Deep nerual network;
D O I
暂无
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
A biological inspired algorithm from human brain known as deep neural network (DNN) containing of multiple hidden layers often occurs vanishing gradient problem due to the saturation characteristic of activation function. Thus, the choice of activation function in DNN is crucial to boost up the DNN recognition performance. Unsaturated activation functions i.e. rectified linear unit is recently proposed to prevent vanishing gradient problem happened during the training process of DNN. In this paper, DNN performance is investigated with three categories of activation functions i.e. saturated, unsaturated and adaptive activation functions. The experimental results showed that the saturation problem of hyperbolic tangent activation function can be solved by adding two trainable parameters in its function. The trainable version of rectified linear unit i.e. parametric rectified linear unit (PReLU) obtained lowest misclassification rate among all types of activation function i.e. 1.6% misclassification rate on MNIST handwritten digit dataset. This is due to the adaptive activation functions allows the network to estimate a better solution by training the activation function parameters during the training process. Therefore, adaptive activation functions improves the generalization of the network to deal with the real-world application.
引用
收藏
页码:686 / 690
页数:5
相关论文
共 50 条
  • [31] Enhancement of Hybrid Deep Neural Network Using Activation Function for EEG Based Emotion Recognition
    Matthew, Jehosheba Margaret
    Mustafa, Masoodhu Banu Noordheen Mohammad
    [J]. TRAITEMENT DU SIGNAL, 2024, 41 (04) : 1991 - 2002
  • [32] Flatness Prediction of Cold Rolled Strip Based on Deep Neural Network with Improved Activation Function
    Liu, Jingyi
    Song, Shuni
    Wang, Jiayi
    Balaiti, Maimutimin
    Song, Nina
    Li, Sen
    [J]. SENSORS, 2022, 22 (02)
  • [33] Adaptive activation function neural net for face recognition
    Talukder, A
    Casasent, D
    [J]. IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 549 - 552
  • [34] An adaptive activation function for multilayer feedforward neural networks
    Yu, CC
    Tang, YC
    Liu, BD
    [J]. 2002 IEEE REGION 10 CONFERENCE ON COMPUTERS, COMMUNICATIONS, CONTROL AND POWER ENGINEERING, VOLS I-III, PROCEEDINGS, 2002, : 645 - 650
  • [35] An adaptive activation function for higher order neural networks
    Xu, SX
    Zhang, M
    [J]. AL 2002: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2002, 2557 : 356 - 362
  • [36] An Adaptive Sigmoidal Activation Function Cascading Neural Networ
    Sharma, Sudhir Kumar
    Chandra, Pravin
    [J]. SOFT COMPUTING MODELS IN INDUSTRIAL AND ENVIRONMENTAL APPLICATIONS, 6TH INTERNATIONAL CONFERENCE SOCO 2011, 2011, 87 : 105 - 116
  • [37] A neural nonlinear adaptive filter with a trainable activation function
    Goh, SL
    Mandic, DP
    Bozic, M
    [J]. 2002 6TH SEMINAR ON NEURAL NETWORK APPLICATIONS IN ELECTRICAL ENGINEERING, PROCEEDINGS, 2002, : 7 - 10
  • [38] Deep Neural Network Using Trainable Activation Functions
    Chung, Hoon
    Lee, Sung Joo
    Park, Jeon Gue
    [J]. 2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 348 - 352
  • [39] Deep Neural Nets with Interpolating Function as Output Activation
    Wang, Bao
    Luo, Xiyang
    Li, Zhen
    Zhu, Wei
    Shi, Zuoqiang
    Osher, Stanley J.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [40] Review of ASIC accelerators for deep neural network
    Machupalli, Raju
    Hossain, Masum
    Mandal, Mrinal
    [J]. MICROPROCESSORS AND MICROSYSTEMS, 2022, 89