A Universal Activation Function for Deep Learning

被引:2
|
作者
Hwang, Seung-Yeon [1 ]
Kim, Jeong-Joon [2 ]
机构
[1] Anyang Univ, Dept Comp Engn, Anyang Si 14028, South Korea
[2] Anyang Univ, Dept ICT Convergence Engn, Anyang Si 14028, South Korea
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2023年 / 75卷 / 02期
基金
新加坡国家研究基金会;
关键词
traditional activation function; Deep learning; activation function; convolutional neural network; benchmark datasets; universal activation function;
D O I
10.32604/cmc.2023.037028
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, deep learning has achieved remarkable results in fields that require human cognitive ability, learning ability, and reasoning ability. Activation functions are very important because they provide the ability of artificial neural networks to learn complex patterns through nonlinearity. Various activation functions are being studied to solve problems such as vanishing gradients and dying nodes that may occur in the deep learning process. However, it takes a lot of time and effort for researchers to use the existing activation function in their research. Therefore, in this paper, we propose a universal activation function (UA) so that researchers can easily create and apply various activation functions and improve the performance of neural networks. UA can generate new types of activation functions as well as functions like traditional activation functions by properly adjusting three hyperparameters. The famous Convolutional Neural Network (CNN) and benchmark dataset were used to evaluate the experimental performance of the UA proposed in this study. We compared the performance of the artificial neural network to which the traditional activation function is applied and the artificial neural network to which the UA is applied. In addition, we evaluated the performance of the new activation function generated by adjusting the hyperparameters of the UA. The experimental performance evaluation results showed that the classification performance of CNNs improved by up to 5% through the UA, although most of them showed similar performance to the
引用
收藏
页码:3553 / 3569
页数:17
相关论文
共 50 条
  • [1] Universal activation function for machine learning
    Yuen, Brosnan
    Hoang, Minh Tu
    Dong, Xiaodai
    Lu, Tao
    [J]. SCIENTIFIC REPORTS, 2021, 11 (01)
  • [2] Universal activation function for machine learning
    Brosnan Yuen
    Minh Tu Hoang
    Xiaodai Dong
    Tao Lu
    [J]. Scientific Reports, 11
  • [3] REPRODUCING ACTIVATION FUNCTION FOR DEEP LEARNING
    Liang, Senwei
    Lyu, Liyao
    Wang, Chunmei
    Yang, Haizhao
    [J]. COMMUNICATIONS IN MATHEMATICAL SCIENCES, 2024, 22 (02) : 285 - 314
  • [4] TeLU: A New Activation Function for Deep Learning
    Mercioni, Marina Adriana
    Holban, Stefan
    [J]. 2020 14TH INTERNATIONAL SYMPOSIUM ON ELECTRONICS AND TELECOMMUNICATIONS (ISETC), 2020, : 32 - 35
  • [5] Smish: A Novel Activation Function for Deep Learning Methods
    Wang, Xueliang
    Ren, Honge
    Wang, Achuan
    [J]. ELECTRONICS, 2022, 11 (04)
  • [6] An Activation Function with Probabilistic Beltrami Coefficient for Deep Learning
    Shimauchi, Hirokazu
    [J]. ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 3, 2022, : 613 - 620
  • [7] ProteinBERT: a universal deep-learning model of protein sequence and function
    Brandes, Nadav
    Ofer, Dan
    Peleg, Yam
    Rappoport, Nadav
    Linial, Michal
    [J]. BIOINFORMATICS, 2022, 38 (08) : 2102 - 2110
  • [8] Soft Clipping Mish - A Novel Activation Function for Deep Learning
    Mercioni, Marina Adriana
    Holban, Stefan
    [J]. 2021 4TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMPUTER TECHNOLOGIES (ICICT 2021), 2021, : 13 - 17
  • [9] An Efficient Hardware Architecture for Activation Function in Deep Learning Processor
    Li, Lin
    Zhang, Shengbing
    Wu, Juan
    [J]. 2018 IEEE 3RD INTERNATIONAL CONFERENCE ON IMAGE, VISION AND COMPUTING (ICIVC), 2018, : 911 - 918
  • [10] Parametric RSigELU: a new trainable activation function for deep learning
    Kilicarslan, Serhat
    Celik, Mete
    [J]. NEURAL COMPUTING & APPLICATIONS, 2024, 36 (13): : 7595 - 7607