Parallel Asynchronous Stochastic Variance Reduction for Nonconvex Optimization

被引:0
|
作者
Fang, Cong
Lin, Zhouchen [1 ]
机构
[1] Peking Univ, Sch EECS, Key Lab Machine Percept MOE, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nowadays, asynchronous parallel algorithms have received much attention in the optimization field due to the crucial demands for modern large-scale optimization problems. However, most asynchronous algorithms focus on convex problems. Analysis on nonconvex problems is lacking. For the Asynchronous Stochastic Descent (ASGD) algorithm, the best result from (Lian et al. 2015) can only achieve an asymptotic O(1/epsilon(2)) rate (convergence to the stationary points, namely, parallel to del f(x)parallel to(2) <= c) on nonconvex problems. In this paper, we study Stochastic Variance Reduced Gradient (SVRG) in the asynchronous setting. We propose the Asynchronous Stochastic Variance Reduced Gradient (ASVRG) algorithm for nonconvex finite-sum problems. We develop two schemes for ASVRG, depending on whether the parameters are updated as an atom or not. We prove that both of the two schemes can achieve linear speed up(1)(a non-asymptotic O(n(2/3)/epsilon) rate to the stationary points) for nonconvex problems when the delay parameter tau < n(1/3), where n is the number of training samples. We also establish a non-asymptotic O(n(2/3) tau(1/3)/epsilon) rate (convergence to the stationary points) for our algorithm without assumptions on t. This further demonstrates that even with asynchronous updating, SVRG has less number of Incremental First-order Oracles (IFOs) compared with Stochastic Gradient Descent and Gradient Descent. We also conduct experiments on a shared memory multi-core system to demonstrate the efficiency of our algorithm.
引用
收藏
页码:794 / 800
页数:7
相关论文
共 50 条
  • [41] Asynchronous Byzantine-Robust Stochastic Aggregation with Variance Reduction for Distributed Learning
    Zhu, Zehan
    Huang, Yan
    Zhao, Chengcheng
    Xu, Jinming
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 151 - 158
  • [42] Dimensionality Reduction for Stationary Time Series via Stochastic Nonconvex Optimization
    Chen, Minshuo
    Yang, Lin F.
    Wang, Mengdi
    Zhao, Tuo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [43] Parallel sequential Monte Carlo for stochastic gradient-free nonconvex optimization
    Akyildiz, Omer Deniz
    Crisan, Dan
    Miguez, Joaquin
    STATISTICS AND COMPUTING, 2020, 30 (06) : 1645 - 1663
  • [44] Parallel sequential Monte Carlo for stochastic gradient-free nonconvex optimization
    Ömer Deniz Akyildiz
    Dan Crisan
    Joaquín Míguez
    Statistics and Computing, 2020, 30 : 1645 - 1663
  • [45] A Parallel Decomposition Method for Nonconvex Stochastic Multi-Agent Optimization Problems
    Yang, Yang
    Scutari, Gesualdo
    Palomar, Daniel P.
    Pesavento, Marius
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (11) : 2949 - 2964
  • [46] Momentum-Based Variance-Reduced Proximal Stochastic Gradient Method for Composite Nonconvex Stochastic Optimization
    Xu, Yangyang
    Xu, Yibo
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2023, 196 (01) : 266 - 297
  • [47] Momentum-Based Variance-Reduced Proximal Stochastic Gradient Method for Composite Nonconvex Stochastic Optimization
    Yangyang Xu
    Yibo Xu
    Journal of Optimization Theory and Applications, 2023, 196 : 266 - 297
  • [48] Stochastic Anderson Mixing for Nonconvex Stochastic Optimization
    Wei, Fuchao
    Bao, Chenglong
    Liu, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [49] Accelerated Stochastic Variance Reduction for a Class of Convex Optimization Problems
    He, Lulu
    Ye, Jimin
    Jianwei, E.
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2023, 196 (03) : 810 - 828
  • [50] On variance reduction for stochastic smooth convex optimization with multiplicative noise
    Alejandro Jofré
    Philip Thompson
    Mathematical Programming, 2019, 174 : 253 - 292