Optimizing performance of feedforward and convolutional neural networks through dynamic activation functions

被引:1
|
作者
Rane, Chinmay [1 ]
Tyagi, Kanishka [1 ]
Kline, Adrienne [2 ]
Chugh, Tushar [3 ]
Manry, Michael [1 ]
机构
[1] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[2] Northwestern Univ, Ctr Artificial Intelligence, Div Cardiac Surg, Northwestern Med, Chicago, IL 60201 USA
[3] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
关键词
Second order algorithms; Output weight optimization; Orthogonal least squares; Dynamic activation functions; MODEL;
D O I
10.1007/s12065-024-00973-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training algorithms in the domain of deep learning, have led to significant breakthroughs across diverse and subsequent domains including speech, text, images, and video processing. While the research around deeper network architectures, notably exemplified by ResNet's expansive 152-layer structures, has yielded remarkable outcomes, the exploration of computationally simpler shallow Convolutional Neural Networks (CNN) remains an area for further exploration. Activation functions, crucial in introducing non-linearity within neural networks, have driven substantial advancements. In this paper, we delve into hidden layer activations, particularly examining their complex piece-wise linear attributes. Our comprehensive experiments showcase the superior efficacy of these piece-wise linear activations over traditional Rectified Linear Units across various architectures. We propose a novel Adaptive Activation algorithm, AdAct, exhibiting promising performance improvements in diverse CNN and multilayer perceptron configurations, thereby presenting compelling results to support its usage.
引用
收藏
页码:4083 / 4093
页数:11
相关论文
共 50 条
  • [21] A Stochastic Computing Method For Generating Activation Functions in Multilayer Feedforward Neural Networks
    Ersoy, Durmus
    Erkmen, Burcu
    ELECTRICA, 2021, 21 (03): : 376 - +
  • [22] ENSEMBLES OF FEEDFORWARD-DESIGNED CONVOLUTIONAL NEURAL NETWORKS
    Chen, Yueru
    Yang, Yijing
    Wang, Wei
    Kuo, C. -C. Jay
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3796 - 3800
  • [23] Interpretable convolutional neural networks via feedforward design
    Kuo, C-C. Jay
    Zhang, Min
    Li, Siyang
    Duan, Jiali
    Chen, Yueru
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 60 : 346 - 359
  • [24] Optimizing feedforward neural networks for control chart pattern recognition through genetic algorithms
    Guh, RS
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2004, 18 (02) : 75 - 99
  • [25] On the performance of pairings of activation and loss functions in neural networks
    Soares, Rodrigo G. F.
    Pereira, Enieson J. S.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 326 - 333
  • [26] Improving the Performance of Neural Networks with an Ensemble of Activation Functions
    Nandi, Arijit
    Jana, Nanda Dulal
    Das, Swagatam
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [27] Sentiment analysis through critic learning for optimizing convolutional neural networks with rules
    Zhang, Bowen
    Xu, Xiaofei
    Li, Xutao
    Chen, Xiaojun
    Ye, Yunming
    Wang, Zhongjie
    NEUROCOMPUTING, 2019, 356 : 21 - 30
  • [28] Activation Functions and their Impact on the Training and Performance of Convolutional Neural Network Models
    Onwujekwe, Gerald
    Yoon, Victoria
    AMCIS 2020 PROCEEDINGS, 2020,
  • [29] Modeling dynamic feedforward neural networks with VHDL
    Chen, Xi
    Wang, Gaofeng
    Zhou, Wei
    Xu, Jiangfeng
    Zhang, Qinglin
    FIRST INTERNATIONAL MULTI-SYMPOSIUMS ON COMPUTER AND COMPUTATIONAL SCIENCES (IMSCCS 2006), PROCEEDINGS, VOL 1, 2006, : 785 - +
  • [30] FEEDFORWARD NEURAL NETWORKS FOR THE IDENTIFICATION OF DYNAMIC PROCESSES
    THIBAULT, J
    CHEMICAL ENGINEERING COMMUNICATIONS, 1991, 105 : 109 - 128