SWAG: A Novel Neural Network Architecture Leveraging Polynomial Activation Functions for Enhanced Deep Learning Efficiency

被引:0
|
作者
Safaei, Saeid [1 ]
Woods, Zerotti [2 ]
Rasheed, Khaled [1 ,3 ]
Taha, Thiab R. [1 ]
Safaei, Vahid [4 ]
Gutierrez, Juan B. [5 ]
Arabnia, Hamid R. [1 ]
机构
[1] Univ Georgia, Dept Comp Sci, Athens, GA 30602 USA
[2] Johns Hopkins Univ, Appl Phys Lab, Baltimore, MD 21218 USA
[3] Univ Georgia, Inst Artificial Intelligence, Athens, GA 30602 USA
[4] Univ Isfahan, Dept Mech Engn, Esfahan 8174673441, Iran
[5] Univ Texas San Antonio, Dept Math, San Antonio, TX 78249 USA
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Activation functions; factorial coefficient; neural network design; polynomial activation function; MULTILAYER FEEDFORWARD NETWORKS; ALGORITHMS;
D O I
10.1109/ACCESS.2024.3403457
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning techniques have demonstrated significant capabilities across numerous applications, with deep neural networks (DNNs) showing promising results. However, training these networks efficiently, especially when determining the most suitable nonlinear activation functions, remains a significant challenge. While the ReLU activation function has been widely adopted, other hand-designed functions have been proposed. One such approach is the trainable activation functions. This paper introduces a novel neural network design, the SWAG. In this structure, instead of evolving, activation functions consistently form a polynomial basis. Each hidden layer in this architecture comprises k sub-layers that use polynomial activation functions adjusted by a factorial coefficient, followed by a Concatenate layer and a layer employing a linear activation function. Leveraging the Stone-Weierstrass approximation theorem, we demonstrate that utilizing a diverse set of polynomial activation functions allows neural networks to retain universal approximation capabilities. The SWAG algorithm's architecture is then presented, where data normalization is emphasized, and a new optimized version of SWAG is proposed, which reduces the computational challenge of managing higher degrees of input. This optimization harnesses the Taylor series method by utilizing lower-degree terms to compute higher-degree terms efficiently. This paper thus contributes an innovative neural network architecture that optimizes polynomial activation functions, promising more efficient and robust deep learning applications.
引用
收藏
页码:73363 / 73375
页数:13
相关论文
共 50 条
  • [1] Learning algorithm analysis for deep neural network with ReLu activation functions
    Placzek, Stanislaw
    Placzek, Aleksander
    [J]. COMPUTER APPLICATIONS IN ELECTRICAL ENGINEERING (ZKWE'2018), 2018, 19
  • [2] On Neural Network Activation Functions and Optimizers in Relation to Polynomial Regression
    Pomerat, John
    Segev, Aviv
    Datta, Rituparna
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 6183 - 6185
  • [3] A Novel Activation Function of Deep Neural Network
    Xiangyang, Lin
    Xing, Qinghua
    Han, Zhang
    Feng, Chen
    [J]. Scientific Programming, 2023, 2023
  • [4] Deep Neural Network Using Trainable Activation Functions
    Chung, Hoon
    Lee, Sung Joo
    Park, Jeon Gue
    [J]. 2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 348 - 352
  • [5] Learning Activation Functions in Deep (Spline) Neural Networks
    Bohra, Pakshal
    Campos, Joaquim
    Gupta, Harshit
    Aziznejad, Shayan
    Unser, Michael
    [J]. IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2020, 1 : 295 - 309
  • [6] A Novel Visual Field Prediction Using Deep Learning: A Recurrent Neural Network Architecture
    Park, Keunheung
    [J]. INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2019, 60 (09)
  • [7] A New Class of Polynomial Activation Functions of Deep Learning for Precipitation Forecasting
    Wang, Jiachuan
    Chen, Lei
    Ng, Charles Wang Wai
    [J]. WSDM'22: PROCEEDINGS OF THE FIFTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2022, : 1025 - 1035
  • [8] Polynomial Learning Rate Policy with Warm Restart for Deep Neural Network
    Mishra, Purnendu
    Sarawadekar, Kishor
    [J]. PROCEEDINGS OF THE 2019 IEEE REGION 10 CONFERENCE (TENCON 2019): TECHNOLOGY, KNOWLEDGE, AND SOCIETY, 2019, : 2087 - 2092
  • [9] The impact of activation functions on training and performance of a deep neural network
    Marcu, David C.
    Grava, Cristian
    [J]. 2021 16TH INTERNATIONAL CONFERENCE ON ENGINEERING OF MODERN ELECTRIC SYSTEMS (EMES), 2021, : 126 - 129
  • [10] EvoDNN - An Evolutionary Deep Neural Network with Heterogeneous Activation Functions
    Cui, Peiyu
    Shabash, Boris
    Wiese, Kay C.
    [J]. 2019 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2019, : 2362 - 2369