共 50 条
- [1] Train simultaneously, generalize better: Stability of gradient-based minimax learners [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
- [2] Towards stability and optimality in stochastic gradient descent [J]. ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 1290 - 1298
- [3] Stability and Generalization of Decentralized Stochastic Gradient Descent [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9756 - 9764
- [4] Global Convergence and Stability of Stochastic Gradient Descent [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
- [5] Data-Dependent Stability of Stochastic Gradient Descent [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
- [6] Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
- [7] Stability and Generalization of the Decentralized Stochastic Gradient Descent Ascent Algorithm [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
- [9] Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4582 - 4589
- [10] How to train a discriminative front end with stochastic gradient descent and maximum mutual information [J]. 2005 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU), 2005, : 41 - 46