Optimizing performance of feedforward and convolutional neural networks through dynamic activation functions

被引:1
|
作者
Rane, Chinmay [1 ]
Tyagi, Kanishka [1 ]
Kline, Adrienne [2 ]
Chugh, Tushar [3 ]
Manry, Michael [1 ]
机构
[1] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[2] Northwestern Univ, Ctr Artificial Intelligence, Div Cardiac Surg, Northwestern Med, Chicago, IL 60201 USA
[3] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
关键词
Second order algorithms; Output weight optimization; Orthogonal least squares; Dynamic activation functions; MODEL;
D O I
10.1007/s12065-024-00973-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training algorithms in the domain of deep learning, have led to significant breakthroughs across diverse and subsequent domains including speech, text, images, and video processing. While the research around deeper network architectures, notably exemplified by ResNet's expansive 152-layer structures, has yielded remarkable outcomes, the exploration of computationally simpler shallow Convolutional Neural Networks (CNN) remains an area for further exploration. Activation functions, crucial in introducing non-linearity within neural networks, have driven substantial advancements. In this paper, we delve into hidden layer activations, particularly examining their complex piece-wise linear attributes. Our comprehensive experiments showcase the superior efficacy of these piece-wise linear activations over traditional Rectified Linear Units across various architectures. We propose a novel Adaptive Activation algorithm, AdAct, exhibiting promising performance improvements in diverse CNN and multilayer perceptron configurations, thereby presenting compelling results to support its usage.
引用
收藏
页码:4083 / 4093
页数:11
相关论文
共 50 条
  • [1] Novel neuronal activation functions for feedforward neural networks
    Efe, Mehmet Oender
    NEURAL PROCESSING LETTERS, 2008, 28 (02) : 63 - 79
  • [2] Novel Neuronal Activation Functions for Feedforward Neural Networks
    Mehmet Önder Efe
    Neural Processing Letters, 2008, 28 : 63 - 79
  • [3] Piecewise Polynomial Activation Functions for Feedforward Neural Networks
    Lopez-Rubio, Ezequiel
    Ortega-Zamorano, Francisco
    Dominguez, Enrique
    Munoz-Perez, Jose
    NEURAL PROCESSING LETTERS, 2019, 50 (01) : 121 - 147
  • [4] Piecewise Polynomial Activation Functions for Feedforward Neural Networks
    Ezequiel López-Rubio
    Francisco Ortega-Zamorano
    Enrique Domínguez
    José Muñoz-Pérez
    Neural Processing Letters, 2019, 50 : 121 - 147
  • [5] Optimizing nonlinear activation function for convolutional neural networks
    Munender Varshney
    Pravendra Singh
    Signal, Image and Video Processing, 2021, 15 : 1323 - 1330
  • [6] Optimizing nonlinear activation function for convolutional neural networks
    Varshney, Munender
    Singh, Pravendra
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (06) : 1323 - 1330
  • [7] Adaptive activation functions in convolutional neural networks
    Qian, Sheng
    Liu, Hua
    Liu, Cheng
    Wu, Si
    Wong, Hau San
    NEUROCOMPUTING, 2018, 272 : 204 - 212
  • [8] Analysis of Activation Functions for Particle Swarm Optimised Feedforward Neural Networks
    van Wyk, Andrich B.
    Engelbrecht, Andries P.
    2016 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2016, : 423 - 430
  • [9] Feedforward neural networks based on PPS-wavelet activation functions
    Marar, JF
    Filho, ECBC
    Vasconcelos, GC
    II WORKSHOP ON CYBERNETIC VISION, PROCEEDINGS, 1997, : 240 - 245
  • [10] Constructive feedforward neural networks using hermite polynomial activation functions
    Ma, LY
    Khorasani, K
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2005, 16 (04): : 821 - 833