Adaptive activation functions accelerate convergence in deep and physics-informed neural networks

被引:454
|
作者
Jagtap, Ameya D. [1 ]
Kawaguchi, Kenji [2 ]
Karniadakis, George Em [1 ,3 ]
机构
[1] Brown Univ, Div Appl Math, 182 George St, Providence, RI 02912 USA
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[3] Pacific Northwest Natl Lab, Richland, WA 99354 USA
关键词
Machine learning; Bad minima; Inverse problems; Physics-informed neural networks; Partial differential equations; Deep learning benchmarks; MODEL;
D O I
10.1016/j.jcp.2019.109136
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We employ adaptive activation functions for regression in deep and physics-informed neural networks (PINNs) to approximate smooth and discontinuous functions as well as solutions of linear and nonlinear partial differential equations. In particular, we solve the nonlinear Klein-Gordon equation, which has smooth solutions, the nonlinear Burgers equation, which can admit high gradient solutions, and the Helmholtz equation. We introduce a scalable hyper-parameter in the activation function, which can be optimized to achieve best performance of the network as it changes dynamically the topology of the loss function involved in the optimization process. The adaptive activation function has better learning capabilities than the traditional one (fixed activation) as it improves greatly the convergence rate, especially at early training, as well as the solution accuracy. To better understand the learning process, we plot the neural network solution in the frequency domain to examine how the network captures successively different frequency bands present in the solution. We consider both forward problems, where the approximate solutions are obtained, as well as inverse problems, where parameters involved in the governing equation are identified. Our simulation results show that the proposed method is a very simple and effective approach to increase the efficiency, robustness and accuracy of the neural network approximation of nonlinear functions as well as solutions of partial differential equations, especially for forward problems. We theoretically prove that in the proposed method, gradient descent algorithms are not attracted to suboptimal critical points or local minima. Furthermore, the proposed adaptive activation functions are shown to accelerate the minimization process of the loss values in standard deep learning benchmarks using CIFAR-10, CIFAR-100, SVHN, MNIST, KMNIST, Fashion-MNIST, and Semeion datasets with and without data augmentation. (C) 2019 Elsevier Inc. All rights reserved.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] Physics-informed neural networks for periodic flows
    Shah, Smruti
    Anand, N. K.
    [J]. PHYSICS OF FLUIDS, 2024, 36 (07)
  • [32] Physics-Informed Neural Networks with Generalized Residual-Based Adaptive Sampling
    Song, Xiaotian
    Deng, Shuchao
    Fan, Jiahao
    Sun, Yanan
    [J]. ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14863 : 320 - 332
  • [33] Learning of viscosity functions in rarefied gas flows with physics-informed neural networks
    Tucny, Jean-Michel
    Durve, Mihir
    Montessori, Andrea
    Succi, Sauro
    [J]. COMPUTERS & FLUIDS, 2024, 269
  • [34] Numerical analysis of physics-informed neural networks and related models in physics-informed machine learning
    De Ryck, Tim
    Mishra, Siddhartha
    [J]. ACTA NUMERICA, 2024, 33 : 633 - 713
  • [35] STACKED NETWORKS IMPROVE PHYSICS-INFORMED TRAINING: APPLICATIONS TO NEURAL NETWORKS AND DEEP OPERATOR NETWORKS
    Howard, Amanda a.
    Murphy, Sarah h.
    Ahmed, Shady e.
    Stinis, Panos
    [J]. FOUNDATIONS OF DATA SCIENCE, 2024,
  • [36] Physics-informed distribution transformers via molecular dynamics and deep neural networks
    Cai, Difeng
    [J]. JOURNAL OF COMPUTATIONAL PHYSICS, 2022, 468
  • [37] Learning scattering waves via coupling physics-informed neural networks and their convergence analysis
    Zhang, Rui
    Gao, Yu
    [J]. JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2024, 446
  • [38] Physics-informed neural networks: A deep learning framework for solving the vibrational problems
    Wang, Xusheng
    Zhang, Liang
    [J]. ADVANCES IN NANO RESEARCH, 2021, 11 (05) : 495 - 519
  • [39] Variable separated physics-informed neural networks based on adaptive weighted loss functions for blood flow model
    Liu, Youqiong
    Cai, Li
    Chen, Yaping
    Ma, Pengfei
    Zhong, Qian
    [J]. COMPUTERS & MATHEMATICS WITH APPLICATIONS, 2024, 153 : 108 - 122
  • [40] Deep Lyapunov-Based Physics-Informed Neural Networks (DeLb-PINN) for Adaptive Control Design
    Hart, Rebecca G.
    Patil, Omkar Sudhir
    Griffis, Emily J.
    Dixon, Warren E.
    [J]. 2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 1511 - 1516