Elastic exponential linear units for convolutional neural networks

被引:32
|
作者
Kim, Daeho [1 ]
Kim, Jinah [2 ]
Kim, Jaeil [1 ,3 ]
机构
[1] Kyungpook Natl Univ, Dept Artificial Intelligence, 80 Daegu Ro, Daegu, South Korea
[2] Korea Inst Ocean Sci & Technol, Marine Disaster Res Ctr, 385 Haeyang Ro, Busan, South Korea
[3] Kyungpook Natl Univ, Sch Comp Sci & Engn, 80 Daehak Ro, Daegu, South Korea
关键词
NOISE;
D O I
10.1016/j.neucom.2020.03.051
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Activation functions play important roles in determining the depth and non-linearity of deep learning models. Since the Rectified Linear Unit (ReLU) was introduced, many modifications, in which noise is intentionally injected, have been proposed to avoid overfitting. Exponential Linear Unit (ELU) and their variants, with trainable parameters, have been proposed to reduce the bias shift effects which are often observed in ReLU-type activation functions. In this paper, we propose a novel activation function, called the Elastic Exponential Linear Unit (EELU), which combines the advantages of both types of activation functions in a generalized form. EELU has an elastic slope in the positive part, and preserves the negative signal by using a small non-zero gradient. We also present a new strategy to insert neuronal noise using a Gaussian distribution in the activation function to improve generalization. We demonstrated how EELU can represent a wider variety of features with random noise than other activation functions, by visualizing the latent features of convolutional neural networks. We evaluated the effectiveness of the EELU approach through extensive experiments with image classification using the CIFAR-10/CIFAR-100, ImageNet, and Tiny ImageNet datasets. Our experimental results show that EELU achieved better generalization performance and improved classification accuracy over conventional activation functions, such as ReLU, ELU, ReLU- and ELU-like variants, Scaled ELU, and Swish. EELU produced performance improvements in image classification using a smaller number of training samples, owing to its noise injection strategy, which allows significant variation in function outputs, including deactivation. © 2020 The Author(s)
引用
收藏
页码:253 / 266
页数:14
相关论文
共 50 条
  • [1] Rectified Exponential Units for Convolutional Neural Networks
    Ying, Yao
    Su, Jianlin
    Shan, Peng
    Miao, Ligang
    Wang, Xiaolian
    Peng, Silong
    [J]. IEEE ACCESS, 2019, 7 : 101633 - 101640
  • [2] Improved Learning in Convolutional Neural Networks with Shifted Exponential Linear Units (ShELUs)
    Grelsson, Bertil
    Felsberg, Michael
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 517 - 522
  • [3] Hyperbolic Linear Units for Deep Convolutional Neural Networks
    Li, Jia
    Xu, Hua
    Deng, Junhui
    Sun, Xiaomin
    [J]. 2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 353 - 359
  • [4] One-Dimensional Convolutional Neural Networks Based on Exponential Linear Units for Bearing Fault Diagnosis
    Kong, Hanyang
    Yang, Qingyu
    Zhang, Zhiqiang
    Nai, Yongqiang
    An, Dou
    Liu, Yibo
    [J]. 2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 1052 - 1057
  • [5] Parametric Exponential Linear Unit for Deep Convolutional Neural Networks
    Trottier, Ludovic
    Giguere, Philippe
    Chaib-draa, Brahim
    [J]. 2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, : 207 - 214
  • [6] Parametric Deformable Exponential Linear Units for deep neural networks
    Cheng, Qishang
    Li, HongLiang
    Wu, Qingbo
    Ma, Lei
    Ngan, King Ngi
    [J]. NEURAL NETWORKS, 2020, 125 : 281 - 289
  • [7] FReLU: Flexible Rectified Linear Units for Improving Convolutional Neural Networks
    Qiu, Suo
    Xu, Xiangmin
    Cai, Bolun
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 1223 - 1228
  • [8] Deep neural networks with Elastic Rectified Linear Units for object recognition
    Jiang, Xiaoheng
    Pang, Yanwei
    Li, Xuelong
    Pan, Jing
    Xie, Yinghong
    [J]. NEUROCOMPUTING, 2018, 275 : 1132 - 1139
  • [9] Understanding and Improving Convolutional Neural Networks via Concatenated Rectified Linear Units
    Shang, Wenling
    Sohn, Kihyuk
    Almeida, Diogo
    Lee, Honglak
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [10] Elastic Adaptively Parametric Compounded Units for Convolutional Neural Network
    Zhang, Changfan
    Xu, Yifu
    Sheng, Zhenwen
    [J]. JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2023, 27 (04) : 576 - 584