共 50 条
- [1] Understanding Approximate Fisher Information for Fast Convergence of Natural Gradient Descent in Wide Neural Networks [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
- [2] Understanding approximate Fisher information for fast convergence of natural gradient descent in wide neural networks* [J]. JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2021, 2021 (12):
- [3] Learning Overparameterized Neural Networks via Stochastic Gradient Descent on Structured Data [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
- [4] Gradient Descent with Early Stopping is Provably Robust to Label Noise for Overparameterized Neural Networks [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 4313 - 4324
- [5] A Convergence Analysis of Gradient Descent on Graph Neural Networks [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
- [6] Convergence of gradient descent for learning linear neural networks [J]. ADVANCES IN CONTINUOUS AND DISCRETE MODELS, 2024, 2024 (01):
- [7] Convergence of Gradient Descent Algorithm for Diagonal Recurrent Neural Networks [J]. 2007 SECOND INTERNATIONAL CONFERENCE ON BIO-INSPIRED COMPUTING: THEORIES AND APPLICATIONS, 2007, : 29 - 31
- [9] Analysis of natural gradient descent for multilayer neural networks [J]. PHYSICAL REVIEW E, 1999, 59 (04): : 4523 - 4532
- [10] Optimization of Graph Neural Networks with Natural Gradient Descent [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 171 - 179