A Variance Reducing Stochastic Proximal Method with Acceleration Techniques

被引:0
|
作者
Lei, Jialin [1 ]
Zhang, Ying [1 ]
Zhang, Zhao [1 ]
机构
[1] Zhejiang Normal Univ, Sch Math Sci, Jinhua 321004, Peoples R China
来源
TSINGHUA SCIENCE AND TECHNOLOGY | 2023年 / 28卷 / 06期
关键词
composite optimization; Variance Reduction (VR); fast Douglas-Rachford (DR) splitting; proximal operator;
D O I
10.26599/TST.2022.9010051
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We consider a fundamental problem in the field of machine learning-structural risk minimization, which can be represented as the average of a large number of smooth component functions plus a simple and convex (but possibly non-smooth) function. In this paper, we propose a novel proximal variance reducing stochastic method building on the introduced Point-SAGA. Our method achieves two proximal operator calculations by combining the fast Douglas-Rachford splitting and refers to the scheme of the FISTA algorithm in the choice of momentum factors. We show that the objective function value converges to the iteration point at the rate of O(1/k) when each loss function is convex and smooth. In addition, we prove that our method achieves a linear convergence rate for strongly convex and smooth loss functions. Experiments demonstrate the effectiveness of the proposed algorithm, especially when the loss function is ill-conditioned with good acceleration.
引用
收藏
页码:999 / 1008
页数:10
相关论文
共 50 条
  • [41] A Stochastic Composite Gradient Method with Incremental Variance Reduction
    Zhang, Junyu
    Xiao, Lin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [42] EXTRAGRADIENT METHOD WITH VARIANCE REDUCTION FOR STOCHASTIC VARIATIONAL INEQUALITIES
    Iusem, A. N.
    Jofre, A.
    Oliveira, R. I.
    Thompson, P.
    SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 686 - 724
  • [43] Tracking the gradients using the Hessian: A new look at variance reducing stochastic methods
    Gower, Robert M.
    Le Roux, Nicolas
    Bach, Francis
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [44] Acceleration techniques for the characteristic method in unstructured meshes
    Santandrea, S
    Sanchez, R
    ANNALS OF NUCLEAR ENERGY, 2002, 29 (04) : 323 - 352
  • [45] Correction to: A Line Search Based Proximal Stochastic Gradient Algorithm with Dynamical Variance Reduction
    Giorgia Franchini
    Federica Porta
    Valeria Ruggiero
    Ilaria Trombini
    Journal of Scientific Computing, 2023, 96
  • [46] VARIABLE METRIC PROXIMAL STOCHASTIC VARIANCE REDUCED GRADIENT METHODS FOR NONCONVEX NONSMOOTH OPTIMIZATION
    Yu, Tengteng
    Liu, Xin-wei
    Dai, Yu-hong
    Sun, J. I. E.
    JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2022, 18 (04) : 2611 - 2631
  • [47] Understanding Nesterov's Acceleration via Proximal Point Method
    Ahn, Kwangjun
    Sra, Suvrit
    2022 SYMPOSIUM ON SIMPLICITY IN ALGORITHMS, SOSA, 2022, : 117 - 130
  • [48] A STOCHASTIC PROXIMAL GRADIENT METHOD FOR LINEAR HYPERSPECTRAL UNMIXING
    Li, Xiuming
    Xu, Fangfang
    Dai, Yu-Hong
    Journal of Applied and Numerical Optimization, 2024, 6 (03): : 323 - 338
  • [49] Kalman filtering techniques for reducing variance of digital speckle displacement measurement noise
    李东晖
    郭立
    Chinese Optics Letters, 2006, (08) : 453 - 456
  • [50] Pre-processing variance reducing techniques in multispectral positron emission tomography
    Yao, R
    Msaki, P
    Lecomte, R
    PHYSICS IN MEDICINE AND BIOLOGY, 1997, 42 (11): : 2233 - 2253