Optimizing nonlinear activation function for convolutional neural networks

被引:37
|
作者
Varshney, Munender [1 ]
Singh, Pravendra [1 ]
机构
[1] Indian Inst Technol Kanpur, Dept Comp Sci & Engn, Kanpur, Uttar Pradesh, India
关键词
FReLU; ReLU; CNN; Convolutional neural network; Activation function;
D O I
10.1007/s11760-021-01863-z
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Activation functions play a critical role in the training and performance of the deep convolutional neural networks. Currently, the rectified linear unit (ReLU) is the most commonly used activation function for the deep CNNs. ReLU is a piecewise linear function that will output the input directly if it is positive, otherwise, it will output zero. In this work, we propose a novel approach to generalize the ReLU activation function using multiple learnable slope parameters. These learnable slope parameters are optimized for every channel, which leads to the learning of a more generalized activation function (a variant of ReLU) corresponding to each channel. This activation is named as fully parametric rectified linear unit (FReLU) and trained using an alternate optimization technique by learning one set of parameters, keeping another set of parameters frozen. Our experiments show that the method outperforms ReLU and its other variant activation functions and also generalizes over various tasks such as image classification, object detection and action recognition in videos. The Top-1 classification accuracy of FReLU on ImageNet improves by 3.75% for MobileNet and similar to 2% for ResNet-50 over ReLU. We also provide various analyses for better interpretability of our proposed activation function.
引用
收藏
页码:1323 / 1330
页数:8
相关论文
共 50 条
  • [31] Optimizing Convolutional Neural Networks for Embedded Systems by Means of Neuroevolution
    Badan, Filip
    Sekanina, Lukas
    THEORY AND PRACTICE OF NATURAL COMPUTING, TPNC 2019, 2019, 11934 : 109 - 121
  • [32] Optimizing Memory Efficiency for Deep Convolutional Neural Networks on GPUs
    Li, Chao
    Yang, Yi
    Feng, Min
    Chakradhar, Srimat
    Zhou, Huiyang
    SC '16: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2016, : 633 - 644
  • [33] Data Dropout: Optimizing Training Data for Convolutional Neural Networks
    Wang, Tianyang
    Huan, Jun
    Li, Bo
    2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 39 - 46
  • [34] Performance Optimizing Method for Sparse Convolutional Neural Networks on GPU
    Dong X.
    Liu L.
    Li J.
    Feng X.-B.
    Ruan Jian Xue Bao/Journal of Software, 2020, 31 (09): : 2944 - 2964
  • [35] Convolutional neural networks and the modulation transfer function
    King, Page
    Koshel, R. John
    APPLICATIONS OF MACHINE LEARNING 2023, 2023, 12675
  • [36] Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks
    Li, Ji
    Yuan, Zihao
    Li, Zhe
    Ding, Caiwen
    Ren, Ao
    Qiu, Qinru
    Draper, Jeffrey
    Wang, Yanzhi
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 1230 - 1236
  • [37] Nonlinear System Modeling using Convolutional Neural Networks
    Lopez, Mario
    Yu, Wen
    2017 14TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTING SCIENCE AND AUTOMATIC CONTROL (CCE), 2017,
  • [38] Periodic Function as Activation Function for Neural Networks
    Xu, Ding
    Guan, Yue
    Cai, Ping-ping
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE: TECHNIQUES AND APPLICATIONS, AITA 2016, 2016, : 179 - 183
  • [39] Stochastic Selection of Activation Layers for Convolutional Neural Networks
    Nanni, Loris
    Lumini, Alessandra
    Ghidoni, Stefano
    Maguolo, Gianluca
    SENSORS, 2020, 20 (06)
  • [40] Grad Centroid Activation Mapping for Convolutional Neural Networks
    Lafabregue, Baptiste
    Weber, Jonathan
    Gancarski, Pierre
    Forestier, Germain
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 184 - 191