Adaptive activation functions accelerate convergence in deep and physics-informed neural networks

被引:454
|
作者
Jagtap, Ameya D. [1 ]
Kawaguchi, Kenji [2 ]
Karniadakis, George Em [1 ,3 ]
机构
[1] Brown Univ, Div Appl Math, 182 George St, Providence, RI 02912 USA
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[3] Pacific Northwest Natl Lab, Richland, WA 99354 USA
关键词
Machine learning; Bad minima; Inverse problems; Physics-informed neural networks; Partial differential equations; Deep learning benchmarks; MODEL;
D O I
10.1016/j.jcp.2019.109136
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We employ adaptive activation functions for regression in deep and physics-informed neural networks (PINNs) to approximate smooth and discontinuous functions as well as solutions of linear and nonlinear partial differential equations. In particular, we solve the nonlinear Klein-Gordon equation, which has smooth solutions, the nonlinear Burgers equation, which can admit high gradient solutions, and the Helmholtz equation. We introduce a scalable hyper-parameter in the activation function, which can be optimized to achieve best performance of the network as it changes dynamically the topology of the loss function involved in the optimization process. The adaptive activation function has better learning capabilities than the traditional one (fixed activation) as it improves greatly the convergence rate, especially at early training, as well as the solution accuracy. To better understand the learning process, we plot the neural network solution in the frequency domain to examine how the network captures successively different frequency bands present in the solution. We consider both forward problems, where the approximate solutions are obtained, as well as inverse problems, where parameters involved in the governing equation are identified. Our simulation results show that the proposed method is a very simple and effective approach to increase the efficiency, robustness and accuracy of the neural network approximation of nonlinear functions as well as solutions of partial differential equations, especially for forward problems. We theoretically prove that in the proposed method, gradient descent algorithms are not attracted to suboptimal critical points or local minima. Furthermore, the proposed adaptive activation functions are shown to accelerate the minimization process of the loss values in standard deep learning benchmarks using CIFAR-10, CIFAR-100, SVHN, MNIST, KMNIST, Fashion-MNIST, and Semeion datasets with and without data augmentation. (C) 2019 Elsevier Inc. All rights reserved.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Locally adaptive activation functions with slope recovery for deep and physics-informed neural networks
    Jagtap, Ameya D.
    Kawaguchi, Kenji
    Karniadakis, George Em
    [J]. PROCEEDINGS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2020, 476 (2239):
  • [2] The Role of Adaptive Activation Functions in Fractional Physics-Informed Neural Networks
    Coelho, C.
    Costa, M. Fernanda P.
    Ferras, L. L.
    [J]. INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS 2022, ICNAAM-2022, 2024, 3094
  • [3] Learning Specialized Activation Functions for Physics-Informed Neural Networks
    Wang, Honghui
    Lu, Lu
    Song, Shiji
    Huang, Gao
    [J]. COMMUNICATIONS IN COMPUTATIONAL PHYSICS, 2023, 34 (04) : 869 - 906
  • [4] Self-adaptive physics-informed neural networks
    McClenny, Levi D.
    Braga-Neto, Ulisses M.
    [J]. JOURNAL OF COMPUTATIONAL PHYSICS, 2023, 474
  • [5] Adaptive task decomposition physics-informed neural networks
    Yang, Jianchuan
    Liu, Xuanqi
    Diao, Yu
    Chen, Xi
    Hu, Haikuo
    [J]. COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2024, 418
  • [6] Physics-Informed Neural Networks for Cardiac Activation Mapping
    Costabal, Francisco Sahli
    Yang, Yibo
    Perdikaris, Paris
    Hurtado, Daniel E.
    Kuhl, Ellen
    [J]. FRONTIERS IN PHYSICS, 2020, 8
  • [7] An Improved Method for Physics-Informed Neural Networks That Accelerates Convergence
    Yan, Liangliang
    Zhou, You
    Liu, Huan
    Liu, Lingqi
    [J]. IEEE ACCESS, 2024, 12 : 23943 - 23953
  • [8] Exact imposition of boundary conditions with distance functions in physics-informed deep neural networks
    Sukumar, N.
    Srivastava, Ankit
    [J]. COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2022, 389
  • [9] Physics-informed neural networks with adaptive localized artificial viscosity
    Coutinho, Emilio Jose Rocha
    Dall'Aqua, Marcelo
    McClenny, Levi
    Zhong, Ming
    Braga-Neto, Ulisses
    Gildin, Eduardo
    [J]. JOURNAL OF COMPUTATIONAL PHYSICS, 2023, 489
  • [10] Physics-Informed Deep Neural Networks for Transient Electromagnetic Analysis
    Noakoasteen, Oameed
    Wang, Shu
    Peng, Zhen
    Christodoulou, Christos
    [J]. IEEE OPEN JOURNAL OF ANTENNAS AND PROPAGATION, 2020, 1 (01): : 404 - 412