Locally adaptive activation functions with slope recovery for deep and physics-informed neural networks

被引:219
|
作者
Jagtap, Ameya D. [1 ]
Kawaguchi, Kenji [2 ]
Karniadakis, George Em [1 ,3 ]
机构
[1] Brown Univ, Div Appl Math, 182 George St, Providence, RI 02912 USA
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[3] Pacific Northwest Natl Lab, Richland, WA 99354 USA
关键词
physics-informed neural networks; machine learning; bad minima; stochastic gradients; accelerated training; deep learning benchmarks;
D O I
10.1098/rspa.2020.0334
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We propose two approaches of locally adaptive activation functions namely, layer-wise and neuron-wise locally adaptive activation functions, which improve the performance of deep and physics-informed neural networks. The local adaptation of activation function is achieved by introducing a scalable parameter in each layer (layer-wise) and for every neuron (neuron-wise) separately, and then optimizing it using a variant of stochastic gradient descent algorithm. In order to further increase the training speed, an activation slope-based slope recovery term is added in the loss function, which further accelerates convergence, thereby reducing the training cost. On the theoretical side, we prove that in the proposed method, the gradient descent algorithms are not attracted to sub-optimal critical points or local minima under practical conditions on the initialization and learning rate, and that the gradient dynamics of the proposed method is not achievable by base methods with any (adaptive) learning rates. We further show that the adaptive activation methods accelerate the convergence by implicitly multiplying conditioning matrices to the gradient of the base method without any explicit computation of the conditioning matrix and the matrix-vector product. The different adaptive activation functions are shown to induce different implicit conditioning matrices. Furthermore, the proposed methods with the slope recovery are shown to accelerate the training process.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Physics-Informed Neural Networks for shell structures
    Bastek, Jan-Hendrik
    Kochmann, Dennis M.
    EUROPEAN JOURNAL OF MECHANICS A-SOLIDS, 2023, 97
  • [32] fPINNs: FRACTIONAL PHYSICS-INFORMED NEURAL NETWORKS
    Pang, Guofei
    Lu, Lu
    Karniadakis, George E. M.
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2019, 41 (04): : A2603 - A2626
  • [33] Physics-informed neural networks for diffraction tomography
    Amirhossein Saba
    Carlo Gigli
    Ahmed B.Ayoub
    Demetri Psaltis
    Advanced Photonics, 2022, 4 (06) : 48 - 59
  • [34] PINNProv: Provenance for Physics-Informed Neural Networks
    de Oliveira, Lyncoln S.
    Kunstmann, Liliane
    Pina, Debora
    de Oliveira, Daniel
    Mattoso, Marta
    2023 INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE AND HIGH PERFORMANCE COMPUTING WORKSHOPS, SBAC-PADW, 2023, : 16 - 23
  • [35] Physics-Informed Neural Networks for Power Systems
    Misyris, George S.
    Venzke, Andreas
    Chatzivasileiadis, Spyros
    2020 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2020,
  • [36] ICPINN: Integral conservation physics-informed neural networks based on adaptive activation functions for 3D blood flow simulations
    Liu, Youqiong
    Cai, Li
    Chen, Yaping
    Chen, Qixing
    COMPUTER PHYSICS COMMUNICATIONS, 2025, 311
  • [37] Surface Profile Recovery from Electromagnetic Fields with Physics-Informed Neural Networks
    Chen, Yuxuan
    Wang, Ce
    Hui, Yuan
    Shah, Nirav Vasant
    Spivack, Mark
    REMOTE SENSING, 2024, 16 (22)
  • [38] Exactly conservative physics-informed neural networks and deep operator networks for dynamical systems
    Cardoso-Bihlo, Elsa
    Bihlo, Alex
    NEURAL NETWORKS, 2025, 181
  • [39] Learning of viscosity functions in rarefied gas flows with physics-informed neural networks
    Tucny, Jean-Michel
    Durve, Mihir
    Montessori, Andrea
    Succi, Sauro
    COMPUTERS & FLUIDS, 2024, 269
  • [40] Physics-Informed Neural Networks with Generalized Residual-Based Adaptive Sampling
    Song, Xiaotian
    Deng, Shuchao
    Fan, Jiahao
    Sun, Yanan
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14863 : 320 - 332