共 50 条
- [41] Decentralized Asynchronous Stochastic Gradient Descent: Convergence Rate Analysis [J]. 2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS (SPCOM 2018), 2018, : 402 - 406
- [42] Fast Convergence for Stochastic and Distributed Gradient Descent in the Interpolation Limit [J]. 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 1890 - 1894
- [43] Implicit Bias of (Stochastic) Gradient Descent for Rank-1 Linear Neural Network [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
- [44] Stochastic Gradient Descent Combines Second-Order Information for Training Neural Network [J]. ICOMS 2018: 2018 INTERNATIONAL CONFERENCE ON MATHEMATICS AND STATISTICS, 2018, : 69 - 73
- [45] Exponential Convergence Time of Gradient Descent for One-Dimensional Deep Linear Neural Networks [J]. CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
- [47] Calibrated Stochastic Gradient Descent for Convolutional Neural Networks [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9348 - 9355
- [49] Stochastic Gradient Descent-Whale Optimization Algorithm-Based Deep Convolutional Neural Network To Crowd Emotion Understanding [J]. COMPUTER JOURNAL, 2020, 63 (02): : 267 - 282
- [50] Universality of gradient descent neural network training [J]. NEURAL NETWORKS, 2022, 150 : 259 - 273