Distributed stochastic gradient tracking methods with momentum acceleration for non-convex optimization

被引:0
|
作者
Juan Gao
Xin-Wei Liu
Yu-Hong Dai
Yakui Huang
Junhua Gu
机构
[1] Hebei University of Technology,School of Artificial Intelligence
[2] Hebei University of Technology,Institute of Mathematics
[3] Chinese Academy of Sciences,LSEC, ICMSEC, Academy of Mathematics and Systems Science
[4] University of Chinese Academy of Sciences,School of Mathematical Sciences
关键词
Distributed non-convex optimization; Machine learning; Momentum methods; Optimization algorithms; Convergence rate;
D O I
暂无
中图分类号
学科分类号
摘要
We consider a distributed non-convex optimization problem of minimizing the sum of all local cost functions over a network of agents. This problem often appears in large-scale distributed machine learning, known as non-convex empirical risk minimization. In this paper, we propose two accelerated algorithms, named DSGT-HB and DSGT-NAG, which combine the distributed stochastic gradient tracking (DSGT) method with momentum accelerated techniques. Under appropriate assumptions, we prove that both algorithms sublinearly converge to a neighborhood of a first-order stationary point of the distributed non-convex optimization. Moreover, we derive the conditions under which DSGT-HB and DSGT-NAG achieve a network-independent linear speedup. Numerical experiments for a distributed non-convex logistic regression problem on real data sets and a deep neural network on the MNIST database show the superiorities of DSGT-HB and DSGT-NAG compared with DSGT.
引用
收藏
页码:531 / 572
页数:41
相关论文
共 50 条
  • [1] Distributed stochastic gradient tracking methods with momentum acceleration for non-convex optimization
    Gao, Juan
    Liu, Xin-Wei
    Dai, Yu-Hong
    Huang, Yakui
    Gu, Junhua
    [J]. COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2023, 84 (02) : 531 - 572
  • [2] Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization
    Jiang, Xia
    Zeng, Xianlin
    Sun, Jian
    Chen, Jie
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) : 5310 - 5321
  • [3] Gradient Methods for Non-convex Optimization
    Prateek Jain
    [J]. Journal of the Indian Institute of Science, 2019, 99 : 247 - 256
  • [4] Gradient Methods for Non-convex Optimization
    Jain, Prateek
    [J]. JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 247 - 256
  • [5] Simple Stochastic Gradient Methods for Non-Smooth Non-Convex Regularized Optimization
    Metel, Michael R.
    Takeda, Akiko
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [6] mPage: Probabilistic Gradient Estimator With Momentum for Non-Convex Optimization
    Liang, Yuqing
    Su, Hui
    Liu, Jinlan
    Xu, Dongpo
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1375 - 1386
  • [7] Riemannian Stochastic Recursive Momentum Method for non-Convex Optimization
    Han, Andi
    Gao, Junbin
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2505 - 2511
  • [8] Adaptive Stochastic Gradient Descent Method for Convex and Non-Convex Optimization
    Chen, Ruijuan
    Tang, Xiaoquan
    Li, Xiuting
    [J]. FRACTAL AND FRACTIONAL, 2022, 6 (12)
  • [9] Decentralized Gradient-Free Methods for Stochastic Non-smooth Non-convex Optimization
    Lin, Zhenwei
    Xia, Jingfan
    Deng, Qi
    Luo, Luo
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 17477 - 17486
  • [10] Non-Convex Distributed Optimization
    Tatarenko, Tatiana
    Touri, Behrouz
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2017, 62 (08) : 3744 - 3757