An Evaluation of Parametric Activation Functions for Deep Learning

被引:0
|
作者
Godfrey, Luke B. [1 ]
机构
[1] SupplyPike, Fayetteville, AR 72703 USA
关键词
NETWORK;
D O I
10.1109/smc.2019.8913972
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Parametric activation functions, such as PReLU and PELU, are a relatively new subdomain of neural network nonlinearities. In this paper, we present a comparison of these methods across several topologies. We find that parameterizing activation functions in neural networks do not tend to overfit and tend to converge more quickly than non-parametric activations. This is especially important in environments where time and resources are limited, such as in embedded and mobile systems. We also introduce the Bendable Linear Unit (BLU), which synthesizes many useful properties of other activations, including PReLU, ELU, and SELU. Our experiments indicate that parametric activations that can approximate the identity function can autonomously learn to make residual connections in deep networks. BLU outperforms other activations on the CIFAR-10 task when using a topology without explicit residual connections. BLU also achieves the highest predictive accuracy of compared activations on the CIFAR-100 task when training with a time limit.
引用
收藏
页码:3006 / 3011
页数:6
相关论文
共 50 条
  • [1] Empirical Evaluation of Activation Functions and Kernel Initializers on Deep Reinforcement Learning
    Jang, Sooyoung
    Son, Youngsung
    [J]. 2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC): ICT CONVERGENCE LEADING THE AUTONOMOUS FUTURE, 2019, : 1140 - 1142
  • [2] Diverse Activation Functions in Deep Learning
    Wang, Bin
    Li, Tianrui
    Huang, Yanyong
    Luo, Huaishao
    Guo, Dongming
    Horng, Shi-Jinn
    [J]. 2017 12TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND KNOWLEDGE ENGINEERING (IEEE ISKE), 2017,
  • [3] Evolutionary Optimization of Deep Learning Activation Functions
    Bingham, Garrett
    Macke, William
    Miikkulainen, Risto
    [J]. GECCO'20: PROCEEDINGS OF THE 2020 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2020, : 289 - 296
  • [4] Parametric RSigELU: a new trainable activation function for deep learning
    Kilicarslan, Serhat
    Celik, Mete
    [J]. NEURAL COMPUTING & APPLICATIONS, 2024, 36 (13): : 7595 - 7607
  • [5] Learning Activation Functions in Deep (Spline) Neural Networks
    Bohra, Pakshal
    Campos, Joaquim
    Gupta, Harshit
    Aziznejad, Shayan
    Unser, Michael
    [J]. IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2020, 1 : 295 - 309
  • [6] Activation functions in deep learning: A comprehensive survey and benchmark
    Dubey, Shiv Ram
    Singh, Satish Kumar
    Chaudhuri, Bidyut Baran
    [J]. NEUROCOMPUTING, 2022, 503 : 92 - 108
  • [7] Discovering Parametric Activation Functions
    Bingham, Garrett
    Miikkulainen, Risto
    [J]. NEURAL NETWORKS, 2022, 148 : 48 - 65
  • [8] THE STUDY OF ACTIVATION FUNCTIONS IN DEEP LEARNING FOR PEDESTRIAN DETECTION AND TRACKING
    Favorskaya, M. N.
    Andreev, V. V.
    [J]. INTERNATIONAL WORKSHOP ON PHOTOGRAMMETRIC AND COMPUTER VISION TECHNIQUES FOR VIDEO SURVEILLANCE, BIOMETRICS AND BIOMEDICINE, 2019, 42-2 (W12): : 53 - 59
  • [9] A framework of deep reinforcement learning for stock evaluation functions
    Luo, Tai-Li
    Wu, Mu-En
    Chen, Chien-Ming
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2020, 38 (05) : 5639 - 5649
  • [10] Effective Activation Functions for Homomorphic Evaluation of Deep Neural Networks
    Obla, Srinath
    Gong, Xinghan
    Aloufi, Asma
    Hu, Peizhao
    Takabi, Daniel
    [J]. IEEE ACCESS, 2020, 8 : 153098 - 153112