Deep-learning density functionals for gradient descent optimization

被引:4
|
作者
Costa, E. [1 ,2 ]
Scriva, G. [1 ,2 ]
Fazio, R. [3 ,4 ]
Pilati, S. [1 ,2 ]
机构
[1] Univ Camerino, Sch Sci & Technol, Phys Div, I-62032 Camerino, Italy
[2] INFN Sez Perugia, I-06123 Perugia, Italy
[3] Abdus Salam Int Ctr Theoret Phys, Str Costiera 11, I-34151 Trieste, Italy
[4] Univ Napoli Federico II, Dipartimento Fis, Monte S Angelo, I-80126 Naples, Italy
基金
欧盟地平线“2020”;
关键词
ANDERSON LOCALIZATION; DIFFUSION; ABSENCE;
D O I
10.1103/PhysRevE.106.045309
中图分类号
O35 [流体力学]; O53 [等离子体物理学];
学科分类号
070204 ; 080103 ; 080704 ;
摘要
Machine-learned regression models represent a promising tool to implement accurate and computationally affordable energy-density functionals to solve quantum many-body problems via density functional theory. However, while they can easily be trained to accurately map ground-state density profiles to the corresponding energies, their functional derivatives often turn out to be too noisy, leading to instabilities in self-consistent iterations and in gradient-based searches of the ground-state density profile. We investigate how these instabilities occur when standard deep neural networks are adopted as regression models, and we show how to avoid them by using an ad hoc convolutional architecture featuring an interchannel averaging layer. The main testbed we consider is a realistic model for noninteracting atoms in optical speckle disorder. With the interchannel average, accurate and systematically improvable ground-state energies and density profiles are obtained via gradient-descent optimization, without instabilities nor violations of the variational principle.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] An efficient, distributed stochastic gradient descent algorithm for deep-learning applications
    Cong, Guojing
    Bhardwaj, Onkar
    Feng, Minwei
    [J]. 2017 46TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING (ICPP), 2017, : 11 - 20
  • [2] Learning Deep Gradient Descent Optimization for Image Deconvolution
    Gong, Dong
    Zhang, Zhen
    Shi, Qinfeng
    van den Hengel, Anton
    Shen, Chunhua
    Zhang, Yanning
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (12) : 5468 - 5482
  • [3] Learning dynamics of gradient descent optimization in deep neural networks
    Wu, Wei
    Jing, Xiaoyuan
    Du, Wencai
    Chen, Guoliang
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2021, 64 (05)
  • [4] Learning dynamics of gradient descent optimization in deep neural networks
    Wei Wu
    Xiaoyuan Jing
    Wencai Du
    Guoliang Chen
    [J]. Science China Information Sciences, 2021, 64
  • [5] Differentially Private Deep Learning with Iterative Gradient Descent Optimization
    Ding, Xiaofeng
    Chen, Lin
    Zhou, Pan
    Jiang, Wenbin
    Jin, Hai
    [J]. ACM/IMS Transactions on Data Science, 2021, 2 (04):
  • [6] Improvement of SPGD by Gradient Descent Optimization Algorithm in Deep Learning
    Zhao, Qingsong
    Hao, Shiqi
    Wang, Yong
    Wang, Lei
    Lin, Zhi
    [J]. 2022 ASIA COMMUNICATIONS AND PHOTONICS CONFERENCE, ACP, 2022, : 469 - 472
  • [7] Learning dynamics of gradient descent optimization in deep neural networks
    Wei WU
    Xiaoyuan JING
    Wencai DU
    Guoliang CHEN
    [J]. Science China(Information Sciences), 2021, 64 (05) : 17 - 31
  • [8] Generalizing projected gradient descent algorithm for massive MIMO detection based on deep-learning
    Yongming, Huang
    Zheng, Wang
    [J]. Dongnan Daxue Xuebao (Ziran Kexue Ban)/Journal of Southeast University (Natural Science Edition), 2024, 54 (04): : 961 - 971
  • [9] Annealed Gradient Descent for Deep Learning
    Pan, Hengyue
    Jiang, Hui
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2015, : 652 - 661
  • [10] Annealed gradient descent for deep learning
    Pan, Hengyue
    Niu, Xin
    Li, RongChun
    Dou, Yong
    Jiang, Hui
    [J]. NEUROCOMPUTING, 2020, 380 (380) : 201 - 211