Enhancement of neural networks with an alternative activation function tanhLU

被引:64
|
作者
Shen, Shui-Long [1 ]
Zhang, Ning [1 ,2 ]
Zhou, Annan [3 ]
Yin, Zhen-Yu [2 ]
机构
[1] Shantou Univ, Coll Engn, Dept Civil & Environm Engn, MOE Key Lab Intelligence Mfg Technol, Shantou 515063, Guangdong, Peoples R China
[2] Hong Kong Polytech Univ, Dept Civil & Environm Engn, Hung Hom, Kowloon, Hong Kong, Peoples R China
[3] RMIT Univ, Sch Engn, Melbourne, Vic 3001, Australia
关键词
Neural networks; Activation function; tanhLUs;
D O I
10.1016/j.eswa.2022.117181
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A novel activation function (referred to as tanhLU) that integrates hyperbolic tangent function (tanh) with a linear unit is proposed as a promising alternative to tanh for neural networks. The tanhLU is inspired by the boundlessness of rectified linear unit (ReLU) and the symmetry of tanh. Three variable parameters in tanhLU controlling activation values and gradients could be preconfigured as constants or adaptively optimized during the training process. The capacity of tanhLU is first investigated by checking the weight gradients in error back propagation. Experiments are conducted to validate the improvement of tanhLUs on five types of neural networks, based on seven benchmark datasets in different domains. tanhLU is then applied to predict the highly nonlinear stress-strain relationship of soils by using the multiscale stress-strain (MSS) dataset. The experiment results indicate that using constant tanhLU leads to apparent improvement on FCNN and LSTM with lower loss and higher accuracy compared with tanh. Adaptive tanhLUs achieved the state-of-the-art performance for multiple deep neural networks in image classification and face recognition.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Regularising neural networks using flexible multivariate activation function
    Solazzi, M
    Uncini, A
    NEURAL NETWORKS, 2004, 17 (02) : 247 - 260
  • [42] High accuracy FPGA activation function implementation for neural networks
    Hajduk, Zbigniew
    NEUROCOMPUTING, 2017, 247 : 59 - 61
  • [43] Activation Function Modulation in Generative Triangular Recurrent Neural Networks
    Sivakumar, Seshadri
    Sivakumar, Shyamala
    IEEE ACCESS, 2022, 10 : 45709 - 45725
  • [44] Wavelet networks as an alternative to neural networks
    Ciuca, I
    Ware, JA
    ETFA '97 - 1997 IEEE 6TH INTERNATIONAL CONFERENCE ON EMERGING TECHNOLOGIES AND FACTORY AUTOMATION PROCEEDINGS, 1997, : 353 - 358
  • [45] AN ALTERNATIVE MODEL FOR NEURAL NETWORKS
    DEALMEIDA, RMC
    IGLESIAS, JR
    PHYSICS LETTERS A, 1990, 146 (05) : 239 - 244
  • [46] Efficient Neural Networks on the Edge with FPGAs by Optimizing an Adaptive Activation Function
    Jiang, Yiyue
    Vaicaitis, Andrius
    Dooley, John
    Leeser, Miriam
    SENSORS, 2024, 24 (06)
  • [47] Self-modeling in Hopfield Neural Networks with Continuous Activation Function
    Zarco, Mario
    Froese, Tom
    8TH ANNUAL INTERNATIONAL CONFERENCE ON BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, BICA 2017 (EIGHTH ANNUAL MEETING OF THE BICA SOCIETY), 2018, 123 : 573 - 578
  • [48] Software reliability prediction using neural networks with linear activation function
    Misra, RB
    Sasatte, PV
    ADVANCED RELIABILITY MODELING, 2004, : 333 - 340
  • [49] A novel activation function for multilayer feed-forward neural networks
    Aboubakar Nasser Samatin Njikam
    Huan Zhao
    Applied Intelligence, 2016, 45 : 75 - 82
  • [50] Quaternionic Hopfield neural networks with twin-multistate activation function
    Kobayashi, Masaki
    NEUROCOMPUTING, 2017, 267 : 304 - 310