Improved Learning in Convolutional Neural Networks with Shifted Exponential Linear Units (ShELUs)

被引:0
|
作者
Grelsson, Bertil [1 ]
Felsberg, Michael [1 ]
机构
[1] Linkoping Univ, Comp Vis Lab, Linkoping, Sweden
基金
瑞典研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Exponential Linear Unit (ELU) has been proven to speed up learning and improve the classification performance over activation functions such as ReLU and Leaky ReLU for convolutional neural networks. The reasons behind the improved behavior are that ELU reduces the bias shift, it saturates for large negative inputs and it is continuously differentiable. However, it remains open whether ELU has the optimal shape and we address the quest for a superior activation function. We use a new formulation to tune a piecewise linear activation function during training, to investigate the above question, and learn the shape of the locally optimal activation function. With this tuned activation function, the classification performance is improved and the resulting, learned activation function shows to be ELU-shaped irrespective if it is initialized as a RELU, LReLU or ELU. Interestingly, the learned activation function does not exactly pass through the origin indicating that a shifted ELU-shaped activation function is preferable. This observation leads us to introduce the Shifted Exponential Linear Unit (ShELU) as a new activation function. Experiments on Cifar-100 show that the classification performance is further improved when using the ShELU activation function in comparison with ELU. The improvement is achieved when learning an individual bias shift for each neuron.
引用
收藏
页码:517 / 522
页数:6
相关论文
共 50 条
  • [1] Elastic exponential linear units for convolutional neural networks
    Kim, Daeho
    Kim, Jinah
    Kim, Jaeil
    [J]. NEUROCOMPUTING, 2020, 406 : 253 - 266
  • [2] Rectified Exponential Units for Convolutional Neural Networks
    Ying, Yao
    Su, Jianlin
    Shan, Peng
    Miao, Ligang
    Wang, Xiaolian
    Peng, Silong
    [J]. IEEE ACCESS, 2019, 7 : 101633 - 101640
  • [3] Hyperbolic Linear Units for Deep Convolutional Neural Networks
    Li, Jia
    Xu, Hua
    Deng, Junhui
    Sun, Xiaomin
    [J]. 2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 353 - 359
  • [4] One-Dimensional Convolutional Neural Networks Based on Exponential Linear Units for Bearing Fault Diagnosis
    Kong, Hanyang
    Yang, Qingyu
    Zhang, Zhiqiang
    Nai, Yongqiang
    An, Dou
    Liu, Yibo
    [J]. 2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 1052 - 1057
  • [5] Parametric Exponential Linear Unit for Deep Convolutional Neural Networks
    Trottier, Ludovic
    Giguere, Philippe
    Chaib-draa, Brahim
    [J]. 2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, : 207 - 214
  • [6] Parametric Deformable Exponential Linear Units for deep neural networks
    Cheng, Qishang
    Li, HongLiang
    Wu, Qingbo
    Ma, Lei
    Ngan, King Ngi
    [J]. NEURAL NETWORKS, 2020, 125 : 281 - 289
  • [7] A New Improved Learning Algorithm for Convolutional Neural Networks
    Yang, Jie
    Zhao, Junhong
    Lu, Lu
    Pan, Tingting
    Jubair, Sidra
    [J]. PROCESSES, 2020, 8 (03)
  • [8] FReLU: Flexible Rectified Linear Units for Improving Convolutional Neural Networks
    Qiu, Suo
    Xu, Xiangmin
    Cai, Bolun
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 1223 - 1228
  • [9] Understanding and Improving Convolutional Neural Networks via Concatenated Rectified Linear Units
    Shang, Wenling
    Sohn, Kihyuk
    Almeida, Diogo
    Lee, Honglak
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [10] AutoTune: Automatically Tuning Convolutional Neural Networks for Improved Transfer Learning
    Basha, S. H. Shabbeer
    Vinakota, Sravan Kumar
    Pulabaigari, Viswanath
    Mukherjee, Snehasis
    Dubey, Shiv Ram
    [J]. NEURAL NETWORKS, 2021, 133 : 112 - 122