KAF plus RSigELU: a nonlinear and kernel-based activation function for deep neural networks

被引:6
|
作者
Kilicarslan, Serhat [1 ]
Celik, Mete [2 ]
机构
[1] Bandirma Onyedi Eylul Univ, Software Engn Dept, Bandirma, Balikesir, Turkey
[2] Erciyes Univ, Dept Comp Engn, TR-38039 Kayseri, Turkey
来源
NEURAL COMPUTING & APPLICATIONS | 2022年 / 34卷 / 16期
关键词
Kernel-based activation function (KAF); KAF plus RSigELUS; KAF plus RSigELUD; CNN; Deep neural network;
D O I
10.1007/s00521-022-07211-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Activation functions (AFs) are the basis for neural network architectures used in real-world problems to accurately model and learn complex relationships between variables. They are preferred to process the input information coming to the network and to produce the corresponding output. The kernel-based activation function (KAF) offers an extended version of ReLU and sigmoid AFs. Therefore, KAF faced with the problems of bias shift originating from the negative region, vanishing gradient, adaptability, flexibility, and neuron death in parameters during the learning process. In this study, hybrid KAF + RSigELUS and KAF + RSigELUD AFs, which are extended versions of KAF, are proposed. In the proposed AFs, the gauss kernel function is used. The proposed KAF + RSigELUS and KAF + RSigELUD AFs are effective in the positive, negative, and linear activation regions. Performance evaluations of them were conducted on the MNIST, Fashion MNIST, CIFAR-10, and SVHN benchmark datasets. The experimental evaluations show that the proposed AFs overcome existing problems and outperformed ReLU, LReLU, ELU, PReLU, and KAF AFs.
引用
收藏
页码:13909 / 13923
页数:15
相关论文
共 50 条
  • [21] Modified kernel-based nonlinear feature extraction
    Dai, G
    Qian, YT
    Jia, S
    [J]. 2004 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL V, PROCEEDINGS: DESIGN AND IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS INDUSTRY TECHNOLOGY TRACKS MACHINE LEARNING FOR SIGNAL PROCESSING MULTIMEDIA SIGNAL PROCESSING SIGNAL PROCESSING FOR EDUCATION, 2004, : 721 - 724
  • [22] The kernel-based nonlinear multivariate grey model
    Ma, Xin
    Liu, Zhi-bin
    [J]. APPLIED MATHEMATICAL MODELLING, 2018, 56 : 217 - 238
  • [23] Neural Decoding with Kernel-Based Metric Learning
    Brockmeier, Austin J.
    Choi, John S.
    Kriminger, Evan G.
    Francis, Joseph T.
    Principe, Jose C.
    [J]. NEURAL COMPUTATION, 2014, 26 (06) : 1080 - 1107
  • [24] Optimizing nonlinear activation function for convolutional neural networks
    Munender Varshney
    Pravendra Singh
    [J]. Signal, Image and Video Processing, 2021, 15 : 1323 - 1330
  • [25] Optimizing nonlinear activation function for convolutional neural networks
    Varshney, Munender
    Singh, Pravendra
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (06) : 1323 - 1330
  • [26] On the Impact of the Activation Function on Deep Neural Networks Training
    Hayou, Soufiane
    Doucet, Arnaud
    Rousseau, Judith
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [27] Pathological brain classification using multiple kernel-based deep convolutional neural network
    Dora, Lingraj
    Agrawal, Sanjay
    Panda, Rutuparna
    Pachori, Ram Bilas
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 36 (2): : 747 - 756
  • [28] Pathological brain classification using multiple kernel-based deep convolutional neural network
    Dora, Lingraj
    Agrawal, Sanjay
    Panda, Rutuparna
    Pachori, Ram Bilas
    [J]. Neural Computing and Applications, 2024, 36 (02) : 747 - 756
  • [29] Pathological brain classification using multiple kernel-based deep convolutional neural network
    Lingraj Dora
    Sanjay Agrawal
    Rutuparna Panda
    Ram Bilas Pachori
    [J]. Neural Computing and Applications, 2024, 36 : 747 - 756
  • [30] Kernel-Based Smoothness Analysis of Residual Networks
    Tirer, Tom
    Bruna, Joan
    Giryes, Raja
    [J]. MATHEMATICAL AND SCIENTIFIC MACHINE LEARNING, VOL 145, 2021, 145 : 921 - 954