Deep neural networks with a set of node-wise varying activation functions

被引:5
|
作者
Jang, Jinhyeok [1 ]
Cho, Hyunjoong [2 ]
Kim, Jaehong [1 ]
Lee, Jaeyeon [1 ]
Yang, Seungjoon [2 ]
机构
[1] Elect & Telecommun Res Inst ETRI, Daejeon, South Korea
[2] Ulsan Natl Inst Sci & Technol UNIST, Sch Elect & Comp Engn, Ulsan, South Korea
关键词
Deep network; Principal component analysis; Pruning; Varying activation;
D O I
10.1016/j.neunet.2020.03.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we present deep neural networks with a set of node-wise varying activation functions. The feature-learning abilities of the nodes are affected by the selected activation functions, where the nodes with smaller indices become increasingly more sensitive during training. As a result, the features learned by the nodes are sorted by the node indices in order of their importance such that more sensitive nodes are related to more important features. The proposed networks learn input features but also the importance of the features. Nodes with lower importance in the proposed networks can be pruned to reduce the complexity of the networks, and the pruned networks can be retrained without incurring performance losses. We validated the feature-sorting property of the proposed method using both shallow and deep networks as well as deep networks transferred from existing networks. (c) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页码:118 / 131
页数:14
相关论文
共 50 条
  • [31] Wavelets as activation functions in Neural Networks
    Herrera, Oscar
    Priego, Belem
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (05) : 4345 - 4355
  • [32] Learning deep neural networks for node classification
    Li, Bentian
    Pi, Dechang
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 137 : 324 - 334
  • [33] Activation Ensembles for Deep Neural Networks
    Klabjan, Diego
    Harmon, Mark
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 206 - 214
  • [34] Layer-Wise Weight Decay for Deep Neural Networks
    Ishii, Masato
    Sato, Atsushi
    IMAGE AND VIDEO TECHNOLOGY (PSIVT 2017), 2018, 10749 : 276 - 289
  • [35] Batch-wise Regularization of Deep Neural Networks for Interpretability
    Burkart, Nadia
    Faller, Philipp M.
    Peinsipp, Elisabeth
    Huber, Marco F.
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTISENSOR FUSION AND INTEGRATION FOR INTELLIGENT SYSTEMS (MFI), 2020, : 216 - 222
  • [36] Stochastic Layer-Wise Precision in Deep Neural Networks
    Lacey, Griffin
    Taylor, Graham W.
    Areibi, Shawki
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2018, : 663 - 672
  • [37] On extended dissipativity analysis for neural networks with time-varying delay and general activation functions
    Xin Wang
    Kun She
    Shouming Zhong
    Jun Cheng
    Advances in Difference Equations, 2016
  • [38] On extended dissipativity analysis for neural networks with time-varying delay and general activation functions
    Wang, Xin
    She, Kun
    Zhong, Shouming
    Cheng, Jun
    ADVANCES IN DIFFERENCE EQUATIONS, 2016,
  • [39] Exponential Stabilization of Neural Networks With Various Activation Functions and Mixed Time-Varying Delays
    Phat, V. N.
    Trinh, H.
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2010, 21 (07): : 1180 - 1184
  • [40] On exponential stability analysis for neural networks with time-varying delays and general activation functions
    Wang, Yijing
    Yang, Cuili
    Zuo, Zhiqiang
    COMMUNICATIONS IN NONLINEAR SCIENCE AND NUMERICAL SIMULATION, 2012, 17 (03) : 1447 - 1459