Stochastic Gradient Descent on a Tree: an Adaptive and Robust Approach to Stochastic Convex Optimization

被引:0
|
作者
Vakili, Sattar [1 ]
Salgia, Sudeep [2 ]
Zhao, Qing [2 ]
机构
[1] Prowlerio, Cambridge, England
[2] Cornell Univ, Sch Elect & Comp Engn, Ithaca, NY 14853 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/allerton.2019.8919740
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Online minimization of an unknown convex function over the interval [0, 1] is considered under first-order stochastic bandit feedback, which returns a random realization of the gradient of the function at each query point. Without knowing the distribution of the random gradients, a learning algorithm sequentially chooses query points with the objective of minimizing regret defined as the expected cumulative loss of the function values at the query points in excess to the minimum value of the function. An approach based on devising a biased random walk on an infinite-depth binary tree constructed through successive partitioning of the domain of the function is developed. Each move of the random walk is guided by a sequential test based on confidence bounds on the empirical mean constructed using the law of the iterated logarithm. With no tuning parameters, this learning algorithm is robust to heavy-tailed noise with infinite variance and adaptive to unknown function characteristics (specifically, convex, strongly convex, and nonsmooth). It achieves the corresponding optimal regret orders (up to a root log T or a log log T factor) in each class of functions and offers better or matching regret orders than the classical stochastic gradient descent approach which requires the knowledge of the function characteristics for tuning the sequence of step-sizes.
引用
收藏
页码:432 / 438
页数:7
相关论文
共 50 条
  • [1] Adaptive Stochastic Gradient Descent Method for Convex and Non-Convex Optimization
    Chen, Ruijuan
    Tang, Xiaoquan
    Li, Xiuting
    FRACTAL AND FRACTIONAL, 2022, 6 (12)
  • [2] ON THE PRIVACY OF NOISY STOCHASTIC GRADIENT DESCENT FOR CONVEX OPTIMIZATION
    Altschuler, Jason M.
    Bok, Jinho
    Talwar, Kunal
    SIAM JOURNAL ON COMPUTING, 2024, 53 (04) : 969 - 1001
  • [3] CONTROLLING STOCHASTIC GRADIENT DESCENT USING STOCHASTIC APPROXIMATION FOR ROBUST DISTRIBUTED OPTIMIZATION
    Jain, Adit
    Krishnamurthy, Vikram
    NUMERICAL ALGEBRA CONTROL AND OPTIMIZATION, 2024,
  • [4] Robust Pose Graph Optimization Using Stochastic Gradient Descent
    Wang, John
    Olson, Edwin
    2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2014, : 4284 - 4289
  • [5] Adaptive Sampling for Incremental Optimization Using Stochastic Gradient Descent
    Papa, Guillaume
    Bianchi, Pascal
    Clemencon, Stephan
    ALGORITHMIC LEARNING THEORY, ALT 2015, 2015, 9355 : 317 - 331
  • [6] Optimization by Adaptive Stochastic Descent
    Kerr, Cliff C.
    Dura-Bernal, Salvador
    Smolinski, Tomasz G.
    Chadderdon, George L.
    Wilson, David P.
    PLOS ONE, 2018, 13 (03):
  • [7] A Stochastic Gradient Descent Approach for Stochastic Optimal Control
    Archibald, Richard
    Bao, Feng
    Yong, Jiongmin
    EAST ASIAN JOURNAL ON APPLIED MATHEMATICS, 2020, 10 (04) : 635 - 658
  • [8] Efficient Smooth Non-Convex Stochastic Compositional Optimization via Stochastic Recursive Gradient Descent
    Hu, Wenqing
    Li, Chris Junchi
    Lian, Xiangru
    Liu, Ji
    Yuan, Huizhuo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses
    Bassily, Raef
    Feldman, Vitaly
    Guzman, Cristobal
    Talwar, Kunal
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [10] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    arXiv, 2019,