Locally adaptive activation functions with slope recovery for deep and physics-informed neural networks

被引:219
|
作者
Jagtap, Ameya D. [1 ]
Kawaguchi, Kenji [2 ]
Karniadakis, George Em [1 ,3 ]
机构
[1] Brown Univ, Div Appl Math, 182 George St, Providence, RI 02912 USA
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[3] Pacific Northwest Natl Lab, Richland, WA 99354 USA
关键词
physics-informed neural networks; machine learning; bad minima; stochastic gradients; accelerated training; deep learning benchmarks;
D O I
10.1098/rspa.2020.0334
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We propose two approaches of locally adaptive activation functions namely, layer-wise and neuron-wise locally adaptive activation functions, which improve the performance of deep and physics-informed neural networks. The local adaptation of activation function is achieved by introducing a scalable parameter in each layer (layer-wise) and for every neuron (neuron-wise) separately, and then optimizing it using a variant of stochastic gradient descent algorithm. In order to further increase the training speed, an activation slope-based slope recovery term is added in the loss function, which further accelerates convergence, thereby reducing the training cost. On the theoretical side, we prove that in the proposed method, the gradient descent algorithms are not attracted to sub-optimal critical points or local minima under practical conditions on the initialization and learning rate, and that the gradient dynamics of the proposed method is not achievable by base methods with any (adaptive) learning rates. We further show that the adaptive activation methods accelerate the convergence by implicitly multiplying conditioning matrices to the gradient of the base method without any explicit computation of the conditioning matrix and the matrix-vector product. The different adaptive activation functions are shown to induce different implicit conditioning matrices. Furthermore, the proposed methods with the slope recovery are shown to accelerate the training process.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Adaptive activation functions accelerate convergence in deep and physics-informed neural networks
    Jagtap, Ameya D.
    Kawaguchi, Kenji
    Karniadakis, George Em
    JOURNAL OF COMPUTATIONAL PHYSICS, 2020, 404 (404)
  • [2] The Role of Adaptive Activation Functions in Fractional Physics-Informed Neural Networks
    Coelho, C.
    Costa, M. Fernanda P.
    Ferras, L. L.
    INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS 2022, ICNAAM-2022, 2024, 3094
  • [3] Simple yet effective adaptive activation functions for physics-informed neural networks
    Zhang, Jun
    Ding, Chensen
    COMPUTER PHYSICS COMMUNICATIONS, 2025, 307
  • [4] Learning Specialized Activation Functions for Physics-Informed Neural Networks
    Wang, Honghui
    Lu, Lu
    Song, Shiji
    Huang, Gao
    COMMUNICATIONS IN COMPUTATIONAL PHYSICS, 2023, 34 (04) : 869 - 906
  • [5] Self-adaptive physics-informed neural networks
    McClenny, Levi D.
    Braga-Neto, Ulisses M.
    JOURNAL OF COMPUTATIONAL PHYSICS, 2023, 474
  • [6] Self-Adaptive Physics-Informed Neural Networks
    Texas A&M University, United States
    1600,
  • [7] Adaptive task decomposition physics-informed neural networks
    Yang, Jianchuan
    Liu, Xuanqi
    Diao, Yu
    Chen, Xi
    Hu, Haikuo
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2024, 418
  • [8] Physics-Informed Neural Networks for Cardiac Activation Mapping
    Costabal, Francisco Sahli
    Yang, Yibo
    Perdikaris, Paris
    Hurtado, Daniel E.
    Kuhl, Ellen
    FRONTIERS IN PHYSICS, 2020, 8
  • [9] Exact imposition of boundary conditions with distance functions in physics-informed deep neural networks
    Sukumar, N.
    Srivastava, Ankit
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2022, 389
  • [10] Physics-informed neural networks with adaptive localized artificial viscosity
    Coutinho, Emilio Jose Rocha
    Dall'Aqua, Marcelo
    McClenny, Levi
    Zhong, Ming
    Braga-Neto, Ulisses
    Gildin, Eduardo
    JOURNAL OF COMPUTATIONAL PHYSICS, 2023, 489