共 50 条
- [1] Training Two-Layer ReLU Networks with Gradient Descent is Inconsistent [J]. Journal of Machine Learning Research, 2022, 23
- [2] Implicit Bias of Gradient Descent for Two-layer ReLU and Leaky ReLU Networks on Nearly-orthogonal Data [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
- [5] Annihilation of Spurious Minima in Two-Layer ReLU Networks [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
- [6] Convergence Analysis of Two-layer Neural Networks with ReLU Activation [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
- [8] PLATEAU PHENOMENON IN GRADIENT DESCENT TRAINING OF RELU NETWORKS: EXPLANATION, QUANTIFICATION, AND AVOIDANCE [J]. SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2021, 43 (05): : A3438 - A3468
- [9] Learning One-hidden-layer ReLU Networks via Gradient Descent [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89