Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses

被引:0
|
作者
Bassily, Raef [1 ]
Feldman, Vitaly [2 ]
Guzman, Cristobal [3 ]
Talwar, Kunal [2 ]
机构
[1] Ohio State Univ, Dept Comp Sci & Engn, Columbus, OH 43210 USA
[2] Apple, Cupertino, CA USA
[3] Pontificia Univ Catolica Chile, ANID Millennium Sci Initiat Program, Inst Math & Computat Engn, Millennium Nucleus Ctr Discovery Structures Compl, Santiago, Chile
关键词
NOISE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Uniform stability is a notion of algorithmic stability that bounds the worst case change in the model output by the algorithm when a single data point in the dataset is replaced. An influential work of Hardt et al. [20] provides strong upper bounds on the uniform stability of the stochastic gradient descent (SGD) algorithm on sufficiently smooth convex losses. These results led to important progress in understanding of the generalization properties of SGD and several applications to differentially private convex optimization for smooth losses. Our work is the first to address uniform stability of SGD on nonsmooth convex losses. Specifically, we provide sharp upper and lower bounds for several forms of SGD and full-batch GD on arbitrary Lipschitz nonsmooth convex losses. Our lower bounds show that, in the nonsmooth case, (S) GD can be inherently less stable than in the smooth case. On the other hand, our upper bounds show that (S)GD is sufficiently stable for deriving new and useful bounds on generalization error. Most notably, we obtain the first dimension-independent generalization bounds for multi-pass SGD in the nonsmooth case. In addition, our bounds allow us to derive a new algorithm for differentially private nonsmooth stochastic convex optimization with optimal excess population risk. Our algorithm is simpler and more efficient than the best known algorithm for the nonsmooth case [16].
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Learning with Gradient Descent and Weakly Convex Losses
    Richards, Dominic
    Rabbat, Mike
    [J]. 24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [2] A new inexact gradient descent method with applications to nonsmooth convex optimization
    Khanh, Pham Duy
    Mordukhovich, Boris S.
    Tran, Dat Ba
    [J]. OPTIMIZATION METHODS & SOFTWARE, 2024,
  • [3] A New Inexact Gradient Descent Method with Applications to Nonsmooth Convex Optimization
    Khanh, Pham Duy
    Mordukhovich, Boris S.
    Tran, Dat Ba
    [J]. arXiv, 2023,
  • [4] ON THE PRIVACY OF NOISY STOCHASTIC GRADIENT DESCENT FOR CONVEX OPTIMIZATION
    Altschuler, Jason M.
    Bok, Jinho
    Talwar, Kunal
    [J]. SIAM JOURNAL ON COMPUTING, 2024, 53 (04) : 969 - 1001
  • [5] Online stochastic gradient descent on non-convex losses from high-dimensional inference
    Ben Arous, Gerard
    Gheissari, Reza
    Jagannath, Aukosh
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [6] Online stochastic gradient descent on non-convex losses from high-dimensional inference
    Arous, Gerard Ben
    Gheissari, Reza
    Jagannath, Aukosh
    [J]. Journal of Machine Learning Research, 2021, 22
  • [7] Towards stability and optimality in stochastic gradient descent
    Toulis, Panos
    Tran, Dustin
    Airoldi, Edoardo M.
    [J]. ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 1290 - 1298
  • [8] Stability and Generalization of Decentralized Stochastic Gradient Descent
    Sun, Tao
    Li, Dongsheng
    Wang, Bao
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9756 - 9764
  • [9] Global Convergence and Stability of Stochastic Gradient Descent
    Patel, Vivak
    Zhang, Shushu
    Tian, Bowen
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [10] Adaptive Stochastic Gradient Descent Method for Convex and Non-Convex Optimization
    Chen, Ruijuan
    Tang, Xiaoquan
    Li, Xiuting
    [J]. FRACTAL AND FRACTIONAL, 2022, 6 (12)