Optimizing performance of feedforward and convolutional neural networks through dynamic activation functions

被引:1
|
作者
Rane, Chinmay [1 ]
Tyagi, Kanishka [1 ]
Kline, Adrienne [2 ]
Chugh, Tushar [3 ]
Manry, Michael [1 ]
机构
[1] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[2] Northwestern Univ, Ctr Artificial Intelligence, Div Cardiac Surg, Northwestern Med, Chicago, IL 60201 USA
[3] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
关键词
Second order algorithms; Output weight optimization; Orthogonal least squares; Dynamic activation functions; MODEL;
D O I
10.1007/s12065-024-00973-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training algorithms in the domain of deep learning, have led to significant breakthroughs across diverse and subsequent domains including speech, text, images, and video processing. While the research around deeper network architectures, notably exemplified by ResNet's expansive 152-layer structures, has yielded remarkable outcomes, the exploration of computationally simpler shallow Convolutional Neural Networks (CNN) remains an area for further exploration. Activation functions, crucial in introducing non-linearity within neural networks, have driven substantial advancements. In this paper, we delve into hidden layer activations, particularly examining their complex piece-wise linear attributes. Our comprehensive experiments showcase the superior efficacy of these piece-wise linear activations over traditional Rectified Linear Units across various architectures. We propose a novel Adaptive Activation algorithm, AdAct, exhibiting promising performance improvements in diverse CNN and multilayer perceptron configurations, thereby presenting compelling results to support its usage.
引用
收藏
页码:4083 / 4093
页数:11
相关论文
共 50 条
  • [31] Emergence of Shape Bias in Convolutional Neural Networks through Activation Sparsity
    Li, Tianqin
    Wen, Ziqi
    Li, Yangfan
    Lee, Tai Sing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [32] GPU-based Empirical Evaluation of Activation Functions in Convolutional Neural Networks
    Zaheer, Raniah
    Shaziya, Humera
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON INVENTIVE SYSTEMS AND CONTROL (ICISC 2018), 2018, : 769 - 773
  • [33] Rethinking the Role of Activation Functions in Deep Convolutional Neural Networks for Image Classification
    Zheng, Qinghe
    Yang, Mingqiang
    Tian, Xinyu
    Wang, Xiaochen
    Wang, Deqiang
    ENGINEERING LETTERS, 2020, 28 (01) : 80 - 92
  • [34] Approximation to a compact set of functions by feedforward neural networks
    Wu, Wei
    Nan, Dong
    Li, Zhengxue
    Long, Jinling
    2007 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-6, 2007, : 1222 - 1225
  • [35] Convolutional Neural Networks With Dynamic Regularization
    Wang, Yi
    Bian, Zhen-Peng
    Hou, Junhui
    Chau, Lap-Pui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) : 2299 - 2304
  • [36] On the activation function and fault tolerance in feedforward neural networks
    Hammadi, NC
    Ito, H
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 1998, E81D (01) : 66 - 72
  • [37] LEARNING IN FEEDFORWARD NEURAL NETWORKS BY IMPROVING THE PERFORMANCE
    GORDON, MB
    PERETO, P
    RODRIGUEZGIRONES, M
    PHYSICA A, 1992, 185 (1-4): : 402 - 410
  • [38] An adaptive activation function for multilayer feedforward neural networks
    Yu, CC
    Tang, YC
    Liu, BD
    2002 IEEE REGION 10 CONFERENCE ON COMPUTERS, COMMUNICATIONS, CONTROL AND POWER ENGINEERING, VOLS I-III, PROCEEDINGS, 2002, : 645 - 650
  • [39] Optimizing Dynamic Neural Networks with Brainstorm
    Cui, Weihao
    Han, Zhenhua
    Ouyang, Lingji
    Wang, Yichuan
    Zheng, Ningxin
    Ma, Lingxiao
    Yang, Yuqing
    Yang, Fan
    Xue, Jilong
    Qiu, Lili
    Zhou, Lidong
    Chen, Quan
    Tan, Haisheng
    Guo, Minyi
    PROCEEDINGS OF THE 17TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, OSDI 2023, 2023, : 797 - 815
  • [40] Optimizing convolutional neural networks for IoT devices: performance and energy efficiency of quantization techniques
    Hernandez, Nicolas
    Almeida, Francisco
    Blanco, Vicente
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (09): : 12686 - 12705