Enhancement of neural networks with an alternative activation function tanhLU

被引:64
|
作者
Shen, Shui-Long [1 ]
Zhang, Ning [1 ,2 ]
Zhou, Annan [3 ]
Yin, Zhen-Yu [2 ]
机构
[1] Shantou Univ, Coll Engn, Dept Civil & Environm Engn, MOE Key Lab Intelligence Mfg Technol, Shantou 515063, Guangdong, Peoples R China
[2] Hong Kong Polytech Univ, Dept Civil & Environm Engn, Hung Hom, Kowloon, Hong Kong, Peoples R China
[3] RMIT Univ, Sch Engn, Melbourne, Vic 3001, Australia
关键词
Neural networks; Activation function; tanhLUs;
D O I
10.1016/j.eswa.2022.117181
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A novel activation function (referred to as tanhLU) that integrates hyperbolic tangent function (tanh) with a linear unit is proposed as a promising alternative to tanh for neural networks. The tanhLU is inspired by the boundlessness of rectified linear unit (ReLU) and the symmetry of tanh. Three variable parameters in tanhLU controlling activation values and gradients could be preconfigured as constants or adaptively optimized during the training process. The capacity of tanhLU is first investigated by checking the weight gradients in error back propagation. Experiments are conducted to validate the improvement of tanhLUs on five types of neural networks, based on seven benchmark datasets in different domains. tanhLU is then applied to predict the highly nonlinear stress-strain relationship of soils by using the multiscale stress-strain (MSS) dataset. The experiment results indicate that using constant tanhLU leads to apparent improvement on FCNN and LSTM with lower loss and higher accuracy compared with tanh. Adaptive tanhLUs achieved the state-of-the-art performance for multiple deep neural networks in image classification and face recognition.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] On the activation function and fault tolerance in feedforward neural networks
    Hammadi, NC
    Ito, H
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 1998, E81D (01) : 66 - 72
  • [22] An adaptive activation function for higher order neural networks
    Xu, SX
    Zhang, M
    AL 2002: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2002, 2557 : 356 - 362
  • [23] A Quantum Activation Function for Neural Networks: Proposal and Implementation
    Kumar, Saurabh
    Dangwal, Siddharth
    Adhikary, Soumik
    Bhowmik, Debanjan
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [24] Square Root Based Activation Function in Neural Networks
    Yang, Xiaoyu
    Chen, Yufei
    Liang, Haiquan
    2018 INTERNATIONAL CONFERENCE ON AUDIO, LANGUAGE AND IMAGE PROCESSING (ICALIP), 2018, : 84 - 89
  • [25] RSigELU: A nonlinear activation function for deep neural networks
    Kilicarslan, Serhat
    Celik, Mete
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 174 (174)
  • [26] On the Impact of the Activation Function on Deep Neural Networks Training
    Hayou, Soufiane
    Doucet, Arnaud
    Rousseau, Judith
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [27] Algorithm Research on Improving Activation Function of Convolutional Neural Networks
    Guo, Yanhua
    Sun, Lei
    Zhang, Zhihong
    He, Hong
    PROCEEDINGS OF THE 2019 31ST CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2019), 2019, : 3582 - 3586
  • [28] Activation Function Perturbations in Artificial Neural Networks Effects on Robustness
    Sostre, Justin
    Cahill, Nathan
    Merkel, Cory
    2024 IEEE WESTERN NEW YORK IMAGE AND SIGNAL PROCESSING WORKSHOP, WNYISPW 2024, 2024,
  • [29] Sound synthesis by flexible activation function recurrent neural networks
    Uncini, A
    NEURAL NETS, 2002, 2486 : 168 - 177
  • [30] An Efficient Asymmetric Nonlinear Activation Function for Deep Neural Networks
    Chai, Enhui
    Yu, Wei
    Cui, Tianxiang
    Ren, Jianfeng
    Ding, Shusheng
    SYMMETRY-BASEL, 2022, 14 (05):