A hybrid stochastic optimization framework for composite nonconvex optimization

被引:17
|
作者
Quoc Tran-Dinh [1 ]
Pham, Nhan H. [1 ]
Phan, Dzung T. [2 ]
Nguyen, Lam M. [2 ]
机构
[1] Univ N Carolina, Dept Stat & Operat Res, 318 Hanes Hall, Chapel Hill, NC 27599 USA
[2] IBM Res, Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
基金
美国国家科学基金会;
关键词
Hybrid stochastic estimator; Stochastic optimization algorithm; Oracle complexity; Variance reduction; Composite nonconvex optimization; GRADIENT DESCENT; APPROXIMATION; ALGORITHM;
D O I
10.1007/s10107-020-01583-1
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
We introduce a new approach to develop stochastic optimization algorithms for a class of stochastic composite and possibly nonconvex optimization problems. The main idea is to combine a variance-reduced estimator and an unbiased stochastic one to create a new hybrid estimator which trades-off the variance and bias, and possesses useful properties for developing new algorithms. We first introduce our hybrid estimator and investigate its fundamental properties to form a foundational theory for algorithmic development. Next, we apply our new estimator to develop several variants of stochastic gradient method to solve both expectation and finite-sum composite optimization problems. Our first algorithm can be viewed as a variant of proximal stochastic gradient methods with a single loop and single sample, but can achieve the best-known oracle complexity bound as state-of-the-art double-loop algorithms in the literature. Then, we consider two different variants of our method: adaptive step-size and restarting schemes that have similar theoretical guarantees as in our first algorithm. We also study two mini-batch variants of the proposed methods. In all cases, we achieve the best-known complexity bounds under standard assumptions. We test our algorithms on several numerical examples with real datasets and compare them with many existing methods. Our numerical experiments show that the new algorithms are comparable and, in many cases, outperform their competitors.
引用
收藏
页码:1005 / 1071
页数:67
相关论文
共 50 条
  • [1] A hybrid stochastic optimization framework for composite nonconvex optimization
    Quoc Tran-Dinh
    Nhan H. Pham
    Dzung T. Phan
    Lam M. Nguyen
    [J]. Mathematical Programming, 2022, 191 : 1005 - 1071
  • [2] ProxSARAH: An Efficient Algorithmic Framework for Stochastic Composite Nonconvex Optimization
    Pham, Nhan H.
    Nguyen, Lam M.
    Phan, Dzung T.
    Quoc Tran-Dinh
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [3] Proxsarah: An efficient algorithmic framework for stochastic composite nonconvex optimization
    Pham, Nhan H.
    Nguyen, Lam M.
    Phan, Dzung T.
    Tran-Dinh, Quoc
    [J]. Journal of Machine Learning Research, 2020, 21
  • [4] Mini-batch stochastic approximation methods for nonconvex stochastic composite optimization
    Saeed Ghadimi
    Guanghui Lan
    Hongchao Zhang
    [J]. Mathematical Programming, 2016, 155 : 267 - 305
  • [5] Mini-batch stochastic approximation methods for nonconvex stochastic composite optimization
    Ghadimi, Saeed
    Lan, Guanghui
    Zhang, Hongchao
    [J]. MATHEMATICAL PROGRAMMING, 2016, 155 (1-2) : 267 - 305
  • [6] Unified Algorithm Framework for Nonconvex Stochastic Optimization in Deep Neural Networks
    Zhu, Yini
    Iiduka, Hideaki
    [J]. IEEE ACCESS, 2021, 9 : 143807 - 143823
  • [7] A Variance Reduced Nonconvex Stochastic Optimization framework for Online Kernel Learning
    Pradhan, Hrusikesha
    Rajawat, Ketan
    [J]. 2022 56TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2022, : 1281 - 1285
  • [8] A Generic Acceleration Framework for Stochastic Composite Optimization
    Kulunchakov, Andrei
    Mairal, Julien
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Stochastic Anderson Mixing for Nonconvex Stochastic Optimization
    Wei, Fuchao
    Bao, Chenglong
    Liu, Yang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [10] Stochastic Gauss–Newton algorithm with STORM estimators for nonconvex composite optimization
    Zhaoxin Wang
    Bo Wen
    [J]. Journal of Applied Mathematics and Computing, 2022, 68 : 4621 - 4643