Deep Kronecker neural networks: A general framework for neural networks with adaptive activation functions

被引:90
|
作者
Jagtap, Ameya D. [1 ]
Shin, Yeonjong [1 ]
Kawaguchi, Kenji [2 ]
Karniadakis, George Em [1 ,3 ]
机构
[1] Brown Univ, Div Appl Math, 182 George St, Providence, RI 02912 USA
[2] Harvard Univ, Ctr Math Sci & Applicat, Cambridge, MA 02138 USA
[3] Brown Univ, Sch Engn, Providence, RI 02912 USA
关键词
Deep neural networks; Kronecker product; Rowdy activation functions; Gradient flow dynamics; physics-informed neural networks; Deep learning benchmarks; LEARNING FRAMEWORK;
D O I
10.1016/j.neucom.2021.10.036
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a new type of neural networks, Kronecker neural networks (KNNs), that form a general framework for neural networks with adaptive activation functions. KNNs employ the Kronecker product, which provides an efficient way of constructing a very wide network while keeping the number of parameters low. Our theoretical analysis reveals that under suitable conditions, KNNs induce a faster decay of the loss than that by the feed-forward networks. This is also empirically verified through a set of computational examples. Furthermore, under certain technical assumptions, we establish global convergence of gradient descent for KNNs. As a specific case, we propose the Rowdy activation function that is designed to get rid of any saturation region by injecting sinusoidal fluctuations, which include trainable parameters. The proposed Rowdy activation function can be employed in any neural network architecture like feed-forward neural networks, Recurrent neural networks, Convolutional neural networks etc. The effectiveness of KNNs with Rowdy activation is demonstrated through various computational experiments including function approximation using feed-forward neural networks, solution inference of partial differential equations using the physics-informed neural networks, and standard deep learning benchmark problems using convolutional and fully-connected neural networks. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:165 / 180
页数:16
相关论文
共 50 条
  • [41] Stochastic Neural Networks with Monotonic Activation Functions
    Ravanbakhsh, Siamak
    Poczos, Barnabas
    Schneider, Jeff
    Schuurmans, Dale
    Greiner, Russell
    [J]. ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 809 - 818
  • [42] Learning Activation Functions for Sparse Neural Networks
    Loni, Mohammad
    Mohan, Aditya
    Asadi, Mehdi
    Lindauer, Marius
    [J]. INTERNATIONAL CONFERENCE ON AUTOMATED MACHINE LEARNING, VOL 224, 2023, 224
  • [43] Quantum activation functions for quantum neural networks
    Maronese, Marco
    Destri, Claudio
    Prati, Enrico
    [J]. QUANTUM INFORMATION PROCESSING, 2022, 21 (04)
  • [44] Fractional Adaptation of Activation Functions In Neural Networks
    Zamora Esquivel, Julio
    Cruz Vargas, Jesus Adan
    Lopez-Meyer, Paulo
    Cordourier Maruri, Hector Alfonso
    Camacho Perez, Jose Rodrigo
    Tickoo, Omesh
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7544 - 7550
  • [45] Survey on Activation Functions for Optical Neural Networks
    Destras, Oceane
    Le Beux, Sebastien
    de Magalhaes, Felipe Gohring
    Nicolescu, Gabriela
    [J]. ACM COMPUTING SURVEYS, 2024, 56 (02)
  • [46] Bicomplex Neural Networks with Hypergeometric Activation Functions
    Vieira, Nelson
    [J]. ADVANCES IN APPLIED CLIFFORD ALGEBRAS, 2023, 33 (02)
  • [47] Multistability of Neural Networks with a Class of Activation Functions
    Wang, Lili
    Lu, Wenlian
    Chen, Tianping
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2009, PT 1, PROCEEDINGS, 2009, 5551 : 323 - 332
  • [48] Construction of Activation Functions for Wavelet Neural Networks
    Stepanov, Andrey B.
    [J]. PROCEEDINGS OF 2017 XX IEEE INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND MEASUREMENTS (SCM), 2017, : 397 - 399
  • [49] Comparative analysis of activation functions in neural networks
    Kamalov, Firuz
    Nazir, Amril
    Safaraliev, Murodbek
    Cherukuri, Aswani Kumar
    Zgheib, Rita
    [J]. 2021 28TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS, AND SYSTEMS (IEEE ICECS 2021), 2021,
  • [50] MEDIAN ACTIVATION FUNCTIONS FOR GRAPH NEURAL NETWORKS
    Ruiz, Luana
    Gama, Fernando
    Marques, Antonio G.
    Ribeiro, Alejandro
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 7440 - 7444