Optimizing performance of feedforward and convolutional neural networks through dynamic activation functions

被引:1
|
作者
Rane, Chinmay [1 ]
Tyagi, Kanishka [1 ]
Kline, Adrienne [2 ]
Chugh, Tushar [3 ]
Manry, Michael [1 ]
机构
[1] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[2] Northwestern Univ, Ctr Artificial Intelligence, Div Cardiac Surg, Northwestern Med, Chicago, IL 60201 USA
[3] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
关键词
Second order algorithms; Output weight optimization; Orthogonal least squares; Dynamic activation functions; MODEL;
D O I
10.1007/s12065-024-00973-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training algorithms in the domain of deep learning, have led to significant breakthroughs across diverse and subsequent domains including speech, text, images, and video processing. While the research around deeper network architectures, notably exemplified by ResNet's expansive 152-layer structures, has yielded remarkable outcomes, the exploration of computationally simpler shallow Convolutional Neural Networks (CNN) remains an area for further exploration. Activation functions, crucial in introducing non-linearity within neural networks, have driven substantial advancements. In this paper, we delve into hidden layer activations, particularly examining their complex piece-wise linear attributes. Our comprehensive experiments showcase the superior efficacy of these piece-wise linear activations over traditional Rectified Linear Units across various architectures. We propose a novel Adaptive Activation algorithm, AdAct, exhibiting promising performance improvements in diverse CNN and multilayer perceptron configurations, thereby presenting compelling results to support its usage.
引用
收藏
页码:4083 / 4093
页数:11
相关论文
共 50 条
  • [41] Optimizing Fully Spectral Convolutional Neural Networks on FPGA
    Liu, Shuanglong
    Luk, Wayne
    2020 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2020), 2020, : 39 - 47
  • [42] An Efficient Quantitative Approach for Optimizing Convolutional Neural Networks
    Wang, Yuke
    Feng, Boyuan
    Peng, Xueqiao
    Ding, Yufei
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2050 - 2059
  • [43] Optimizing Convolutional Neural Networks on the Sunway TaihuLight Supercomputer
    Zhao, Wenlai
    Fu, Haohuan
    Fang, Jiarui
    Zheng, Weijie
    Gan, Lin
    Yang, Guangwen
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2018, 15 (01)
  • [44] Optimizing Accelerator on FPGA for Deep Convolutional Neural Networks
    Dong, Yong
    Hu, Wei
    Wang, Yonghao
    Jiao, Qiang
    Chen, Shuang
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 97 - 110
  • [45] PBIL for optimizing inception module in convolutional neural networks
    Garcia-Victoria, Pedro
    Gutierrez-Naranjo, Miguel A.
    Cardenas-Montes, Miguel
    Vasco-Carofilis, Roberto A.
    LOGIC JOURNAL OF THE IGPL, 2023, 31 (02) : 325 - 337
  • [46] Deploying and Optimizing Convolutional Neural Networks on Heterogeneous Architecture
    Jiang, Junning
    Cai, Liang
    Dong, Feng
    Yu, Kehua
    Chen, Ke
    Qu, Wei
    Jiang, Jianfei
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [47] Dynamic tunneling based regularization in feedforward neural networks
    Singh, YP
    RoyChowdhury, P
    ARTIFICIAL INTELLIGENCE, 2001, 131 (1-2) : 55 - 71
  • [48] Evaluation of Activation Functions in Convolutional Neural Networks for Image Classification Based on Homomorphic Encryption
    Jia, Huixue
    Cai, Daomeng
    Huo, Zhilin
    Wang, Cong
    Zhang, Shibin
    Zhang, Shujun
    Li, Xiaoyu
    Yang, Shan
    PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON COMPUTER ENGINEERING AND NETWORKS, VOL III, CENET 2023, 2024, 1127 : 343 - 355
  • [49] A privacy preservation framework for feedforward-designed convolutional neural networks
    Li, De
    Wang, Jinyan
    Li, Qiyu
    Hu, Yuhang
    Li, Xianxian
    NEURAL NETWORKS, 2022, 155 : 14 - 27
  • [50] Deep Convolutional Neural Networks on Cartoon Functions
    Grohs, Philipp
    Wiatowski, Thomas
    Bolcskei, Helmut
    2016 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, 2016, : 1163 - 1167