Deep-learning density functionals for gradient descent optimization

被引:4
|
作者
Costa, E. [1 ,2 ]
Scriva, G. [1 ,2 ]
Fazio, R. [3 ,4 ]
Pilati, S. [1 ,2 ]
机构
[1] Univ Camerino, Sch Sci & Technol, Phys Div, I-62032 Camerino, Italy
[2] INFN Sez Perugia, I-06123 Perugia, Italy
[3] Abdus Salam Int Ctr Theoret Phys, Str Costiera 11, I-34151 Trieste, Italy
[4] Univ Napoli Federico II, Dipartimento Fis, Monte S Angelo, I-80126 Naples, Italy
基金
欧盟地平线“2020”;
关键词
ANDERSON LOCALIZATION; DIFFUSION; ABSENCE;
D O I
10.1103/PhysRevE.106.045309
中图分类号
O35 [流体力学]; O53 [等离子体物理学];
学科分类号
070204 ; 080103 ; 080704 ;
摘要
Machine-learned regression models represent a promising tool to implement accurate and computationally affordable energy-density functionals to solve quantum many-body problems via density functional theory. However, while they can easily be trained to accurately map ground-state density profiles to the corresponding energies, their functional derivatives often turn out to be too noisy, leading to instabilities in self-consistent iterations and in gradient-based searches of the ground-state density profile. We investigate how these instabilities occur when standard deep neural networks are adopted as regression models, and we show how to avoid them by using an ad hoc convolutional architecture featuring an interchannel averaging layer. The main testbed we consider is a realistic model for noninteracting atoms in optical speckle disorder. With the interchannel average, accurate and systematically improvable ground-state energies and density profiles are obtained via gradient-descent optimization, without instabilities nor violations of the variational principle.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] A Limitation of Gradient Descent Learning
    Sum, John
    Leung, Chi-Sing
    Ho, Kevin
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (06) : 2227 - 2232
  • [42] Gradient learning in a classification setting by gradient descent
    Cai, Jia
    Wang, Hongyan
    Zhou, Ding-Xuan
    [J]. JOURNAL OF APPROXIMATION THEORY, 2009, 161 (02) : 674 - 692
  • [43] Learning to Learn Gradient Aggregation by Gradient Descent
    Ji, Jinlong
    Chen, Xuhui
    Wang, Qianlong
    Yu, Lixing
    Li, Pan
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2614 - 2620
  • [44] RETRACTED: Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy (Retracted Article)
    Zhang, Chuanlei
    Yao, Minda
    Chen, Wei
    Zhang, Shanwen
    Chen, Dufeng
    Wu, Yuliang
    [J]. SECURITY AND COMMUNICATION NETWORKS, 2021, 2021
  • [45] NucleoFind: a deep-learning network for interpreting nucleic acid electron density
    Dialpuri, Jordan S.
    Agirre, Jon
    Cowtan, Kathryn D.
    Bond, Paul S.
    [J]. NUCLEIC ACIDS RESEARCH, 2024, 52 (17)
  • [46] Universal materials model of deep-learning density functional theory Hamiltonian
    Wang, Yuxiang
    Li, Yang
    Tang, Zechen
    Li, He
    Yuan, Zilong
    Tao, Honggeng
    Zou, Nianlong
    Bao, Ting
    Liang, Xinghao
    Chen, Zezhou
    Xu, Shanghua
    Bian, Ce
    Xu, Zhiming
    Wang, Chong
    Si, Chen
    Duan, Wenhui
    Xu, Yong
    [J]. SCIENCE BULLETIN, 2024, 69 (16) : 2514 - 2521
  • [47] Solving deep-learning density functional theory via variational autoencoders
    Costa, Emanuele
    Scriva, Giuseppe
    Pilati, Sebastiano
    [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2024, 5 (03):
  • [48] Deep learning-based density functionals empower AI for materials
    Yin, Haoyu
    Chen, Yizhe
    Wang, Xiaonan
    [J]. MATTER, 2022, 5 (08) : 2452 - 2455
  • [49] Deep learning for sea cucumber detection using stochastic gradient descent algorithm
    Zhang, Huaqiang
    Yu, Fusheng
    Sun, Jincheng
    Shen, Xiaoqin
    Li, Kun
    [J]. EUROPEAN JOURNAL OF REMOTE SENSING, 2020, 53 (53-62) : 53 - 62
  • [50] Communication-Efficient Local Stochastic Gradient Descent for Scalable Deep Learning
    Lee, Sunwoo
    Kang, Qiao
    Agrawal, Ankit
    Choudhary, Alok
    Liao, Wei-keng
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 718 - 727