Enhancement of neural networks with an alternative activation function tanhLU

被引:64
|
作者
Shen, Shui-Long [1 ]
Zhang, Ning [1 ,2 ]
Zhou, Annan [3 ]
Yin, Zhen-Yu [2 ]
机构
[1] Shantou Univ, Coll Engn, Dept Civil & Environm Engn, MOE Key Lab Intelligence Mfg Technol, Shantou 515063, Guangdong, Peoples R China
[2] Hong Kong Polytech Univ, Dept Civil & Environm Engn, Hung Hom, Kowloon, Hong Kong, Peoples R China
[3] RMIT Univ, Sch Engn, Melbourne, Vic 3001, Australia
关键词
Neural networks; Activation function; tanhLUs;
D O I
10.1016/j.eswa.2022.117181
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A novel activation function (referred to as tanhLU) that integrates hyperbolic tangent function (tanh) with a linear unit is proposed as a promising alternative to tanh for neural networks. The tanhLU is inspired by the boundlessness of rectified linear unit (ReLU) and the symmetry of tanh. Three variable parameters in tanhLU controlling activation values and gradients could be preconfigured as constants or adaptively optimized during the training process. The capacity of tanhLU is first investigated by checking the weight gradients in error back propagation. Experiments are conducted to validate the improvement of tanhLUs on five types of neural networks, based on seven benchmark datasets in different domains. tanhLU is then applied to predict the highly nonlinear stress-strain relationship of soils by using the multiscale stress-strain (MSS) dataset. The experiment results indicate that using constant tanhLU leads to apparent improvement on FCNN and LSTM with lower loss and higher accuracy compared with tanh. Adaptive tanhLUs achieved the state-of-the-art performance for multiple deep neural networks in image classification and face recognition.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Enhancement of neural networks with an alternative activation function tanhLU
    Shen, Shui-Long
    Zhang, Ning
    Zhou, Annan
    Yin, Zhen-Yu
    Expert Systems with Applications, 2022, 199
  • [2] Periodic Function as Activation Function for Neural Networks
    Xu, Ding
    Guan, Yue
    Cai, Ping-ping
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE: TECHNIQUES AND APPLICATIONS, AITA 2016, 2016, : 179 - 183
  • [3] FUNCTION OF IGG IN THE ENHANCEMENT OF ALTERNATIVE PATHWAY ACTIVATION BY ZYMOSAN
    SCHENKEIN, HA
    RUDDY, S
    FEDERATION PROCEEDINGS, 1980, 39 (03) : 1059 - 1059
  • [4] Neural networks with asymmetric activation function for function approximation
    Gomes, Gecynalda S. da S.
    Ludermir, Teresa B.
    Almeida, Leandro M.
    IJCNN: 2009 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1- 6, 2009, : 2310 - 2317
  • [5] Multistability of neural networks with discontinuous activation function
    Huang, Gan
    Cao, Jinde
    COMMUNICATIONS IN NONLINEAR SCIENCE AND NUMERICAL SIMULATION, 2008, 13 (10) : 2279 - 2289
  • [6] Adaptive Morphing Activation Function for Neural Networks
    Herrera-Alcantara, Oscar
    Arellano-Balderas, Salvador
    FRACTAL AND FRACTIONAL, 2024, 8 (08)
  • [7] Parabola As an Activation Function of Artificial Neural Networks
    Khachumov, M. V.
    Emelyanova, Yu. G.
    SCIENTIFIC AND TECHNICAL INFORMATION PROCESSING, 2024, 51 (05) : 471 - 477
  • [8] Neural networks with adaptive spline activation function
    Campolucci, P
    Capparelli, F
    Guarnieri, S
    Piazza, F
    Uncini, A
    MELECON '96 - 8TH MEDITERRANEAN ELECTROTECHNICAL CONFERENCE, PROCEEDINGS, VOLS I-III: INDUSTRIAL APPLICATIONS IN POWER SYSTEMS, COMPUTER SCIENCE AND TELECOMMUNICATIONS, 1996, : 1442 - 1445
  • [9] Activation function of wavelet chaotic neural networks
    Xu, Yao-Qun
    Sun, Ming
    Guo, Meng-Shu
    PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS, VOLS 1 AND 2, 2006, : 716 - 721
  • [10] DYNAMICS OF NEURAL NETWORKS WITH NONMONOTONE ACTIVATION FUNCTION
    DEFELICE, P
    MARANGI, C
    NARDULLI, G
    PASQUARIELLO, G
    TEDESCO, L
    NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1993, 4 (01) : 1 - 9