Rectified Exponential Units for Convolutional Neural Networks

被引:21
|
作者
Ying, Yao [1 ]
Su, Jianlin [2 ]
Shan, Peng [1 ]
Miao, Ligang [3 ]
Wang, Xiaolian [4 ]
Peng, Silong [1 ,4 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang 110819, Liaoning, Peoples R China
[2] Sun Yat Sen Univ, Sch Math, Guangzhou 510220, Guangdong, Peoples R China
[3] Northeastern Univ, Sch Comp & Commun Engn, Shenyang 110819, Liaoning, Peoples R China
[4] Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China
来源
IEEE ACCESS | 2019年 / 7卷
基金
中国国家自然科学基金;
关键词
Activation function; convolutional neural network; rectified exponential unit; parametric rectified exponential unit;
D O I
10.1109/ACCESS.2019.2928442
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Rectified linear unit (ReLU) plays an important role in today's convolutional neural networks (CNNs). In this paper, we propose a novel activation function called Rectified Exponential Unit (REU). Inspired by two recently proposed activation functions: Exponential Linear Unit (ELU) and Swish, the REU is designed by introducing the advantage of flexible exponent and multiplication function form. Moreover, we propose the Parametric REU (PREU) to increase the expressive power of the REU. The experiments with three classical CNN architectures, LeNet-5, Network in Network, and Residual Network (ResNet) on scale-various benchmarks including Fashion-MNIST, CIFAR10, CIFAR100, and Tiny ImageNet demonstrate that REU and PREU achieve improvement compared with other activation functions. Our results show that our REU has relative error improvements over ReLU of 7.74% and 6.08% on CIFAR-10 and 100 with the ResNet, the improvements of PREU is 9.24% and 9.32%. Finally, we use the different PREU variants in the Residual unit to achieve more stable results.
引用
收藏
页码:101633 / 101640
页数:8
相关论文
共 50 条
  • [1] FReLU: Flexible Rectified Linear Units for Improving Convolutional Neural Networks
    Qiu, Suo
    Xu, Xiangmin
    Cai, Bolun
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 1223 - 1228
  • [2] Elastic exponential linear units for convolutional neural networks
    Kim, Daeho
    Kim, Jinah
    Kim, Jaeil
    [J]. NEUROCOMPUTING, 2020, 406 : 253 - 266
  • [3] Understanding and Improving Convolutional Neural Networks via Concatenated Rectified Linear Units
    Shang, Wenling
    Sohn, Kihyuk
    Almeida, Diogo
    Lee, Honglak
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [4] Improved Learning in Convolutional Neural Networks with Shifted Exponential Linear Units (ShELUs)
    Grelsson, Bertil
    Felsberg, Michael
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 517 - 522
  • [5] Understanding Weight Normalized Deep Neural Networks with Rectified Linear Units
    Xu, Yixi
    Wang, Xiao
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [6] Spam Filtering Using Regularized Neural Networks with Rectified Linear Units
    Barushka, Aliaksandr
    Hajek, Petr
    [J]. AI*IA 2016: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2016, 10037 : 65 - 75
  • [7] Deep neural networks with Elastic Rectified Linear Units for object recognition
    Jiang, Xiaoheng
    Pang, Yanwei
    Li, Xuelong
    Pan, Jing
    Xie, Yinghong
    [J]. NEUROCOMPUTING, 2018, 275 : 1132 - 1139
  • [8] One-Dimensional Convolutional Neural Networks Based on Exponential Linear Units for Bearing Fault Diagnosis
    Kong, Hanyang
    Yang, Qingyu
    Zhang, Zhiqiang
    Nai, Yongqiang
    An, Dou
    Liu, Yibo
    [J]. 2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 1052 - 1057
  • [9] Gradient rectified parameter unit of the fully connected layer in convolutional neural networks
    Zheng, Tianyou
    Wang, Qiang
    Shen, Yue
    Lin, Xiaotian
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 248
  • [10] IMPROVING DEEP NEURAL NETWORKS FOR LVCSR USING RECTIFIED LINEAR UNITS AND DROPOUT
    Dahl, George E.
    Sainath, Tara N.
    Hinton, Geoffrey E.
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 8609 - 8613