Self-adaptive deep neural network: Numerical approximation to functions and PDEs

被引:4
|
作者
Cai, Zhiqiang [1 ]
Chen, Jingshuang [1 ]
Liu, Min [2 ]
机构
[1] Purdue Univ, Dept Math, 150 N Univ St, W Lafayette, IN 47907 USA
[2] Purdue Univ, Sch Mech Engn, 585 Purdue Mall, W Lafayette, IN 47907 USA
基金
美国国家科学基金会;
关键词
Self-adaptivity; Advection-reaction equation; Least-squares approximation; Deep neural network; ReLU activation;
D O I
10.1016/j.jcp.2022.111021
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Designing an optimal deep neural network for a given task is important and challenging in many machine learning applications. To address this issue, we introduce a self-adaptive algorithm: the adaptive network enhancement (ANE) method, written as loops of the form train -> estimate -> enhance. Starting with a small two-layer neural network (NN), the step train is to solve the optimization problem at the current NN; the step estimate is to compute a posteriori estimator/indicators using the solution at the current NN; the step enhance is to add new neurons to the current NN. Novel network enhancement strategies based on the computed estimator/indicators are developed in this paper to determine how many new neurons and when a new layer should be added to the current NN. The ANE method provides a natural process for obtaining a good initialization in training the current NN; in addition, we introduce an advanced procedure on how to initialize newly added neurons for a better approximation. We demonstrate that the ANE method can automatically design a nearly minimal NN for learning functions exhibiting sharp transitional layers as well as discontinuous solutions of hyperbolic partial differential equations. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Self-Adaptive Layer: An Application of Function Approximation Theory to Enhance Convergence Efficiency in Neural Networks
    Chan, Ka-Hou
    Im, Sio-Kei
    Ke, Wei
    [J]. 2020 34TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN 2020), 2020, : 447 - 452
  • [42] Deep Self-Adaptive Hashing for Image Retrieval
    Lin, Qinghong
    Chen, Xiaojun
    Zhang, Qin
    Tian, Shangxuan
    Chen, Yudong
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 1028 - 1037
  • [43] Self-Adaptive Approximate Mobile Deep Learning
    Knez, Timotej
    Machidon, Octavian
    Pejovic, Veljko
    [J]. ELECTRONICS, 2021, 10 (23)
  • [44] Deep belief networks with self-adaptive sparsity
    Qiao, Chen
    Yang, Lan
    Shi, Yan
    Fang, Hanfeng
    Kang, Yanmei
    [J]. APPLIED INTELLIGENCE, 2022, 52 (01) : 237 - 253
  • [45] Study on self-adaptive fuzzy neural networks
    Liu, Fang
    [J]. ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS: WITH ASPECTS OF CONTEMPORARY INTELLIGENT COMPUTING TECHNIQUES, 2007, 2 : 335 - +
  • [46] Designing the Self-Adaptive Fuzzy Neural Networks
    Fang, Liu
    [J]. 2009 INTERNATIONAL JOINT CONFERENCE ON BIOINFORMATICS, SYSTEMS BIOLOGY AND INTELLIGENT COMPUTING, PROCEEDINGS, 2009, : 537 - 540
  • [47] Deep belief networks with self-adaptive sparsity
    Chen Qiao
    Lan Yang
    Yan Shi
    Hanfeng Fang
    Yanmei Kang
    [J]. Applied Intelligence, 2022, 52 : 237 - 253
  • [48] DEEP NETWORK APPROXIMATION FOR SMOOTH FUNCTIONS
    Lu, Jianfeng
    Shen, Zuowei
    Yang, Haizhao
    Zhang, Shijun
    [J]. SIAM JOURNAL ON MATHEMATICAL ANALYSIS, 2021, 53 (05) : 5465 - 5506
  • [49] Neural Network Approximation of Refinable Functions
    Daubechies, Ingrid
    De Vore, Ronald
    Dym, Nadav
    Faigenbaum-Golovin, Shira
    Kovalsky, Shahar Z.
    Lin, Kung-Chin
    Park, Josiah
    Petrova, Guergana
    Sober, Barak
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2023, 69 (01) : 482 - 495
  • [50] Deep Neural Network Approximation Theory
    Elbrachter, Dennis
    Perekrestenko, Dmytro
    Grohs, Philipp
    Boelcskei, Helmut
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2021, 67 (05) : 2581 - 2623