共 50 条
- [1] EXPLORING ONE PASS LEARNING FOR DEEP NEURAL NETWORK TRAINING WITH AVERAGED STOCHASTIC GRADIENT DESCENT [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
- [3] Convergence of Stochastic Gradient Descent in Deep Neural Network [J]. Acta Mathematicae Applicatae Sinica, English Series, 2021, 37 : 126 - 136
- [4] Convergence of Stochastic Gradient Descent in Deep Neural Network [J]. ACTA MATHEMATICAE APPLICATAE SINICA-ENGLISH SERIES, 2021, 37 (01): : 126 - 136
- [6] ASYNCHRONOUS STOCHASTIC GRADIENT DESCENT FOR DNN TRAINING [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 6660 - 6663
- [7] Developing a Loss Prediction-based Asynchronous Stochastic Gradient Descent Algorithm for Distributed Training of Deep Neural Networks [J]. PROCEEDINGS OF THE 49TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2020, 2020,
- [9] Accelerating Asynchronous Stochastic Gradient Descent for Neural Machine Translation [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2991 - 2996
- [10] Universality of gradient descent neural network training [J]. NEURAL NETWORKS, 2022, 150 : 259 - 273