Optimistic Online Mirror Descent for Bridging Stochastic and Adversarial Online Convex Optimization

被引:0
|
作者
Chen, Sijia [1 ,2 ]
Tu, Wei-Wei [1 ,3 ,4 ]
Zhao, Peng [1 ]
Zhang, Lijun [1 ,2 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] 4Paradigm, Beijing, Peoples R China
[4] ChaLearn, Berkeley, CA USA
基金
国家重点研发计划;
关键词
ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastically Extended Adversarial (SEA) model is introduced by Sachs et al. (2022) as an interpolation between stochastic and adversarial online convex optimization. Under the smoothness condition, they demonstrate that the expected regret of optimistic follow-the-regularized-leader (FTRL) depends on the cumulative stochastic variance sigma(2)(1:T) and the cumulative adversarial variation Sigma(2)(1:T) for convex functions. They also provide a slightly weaker bound based on the maximal stochastic variance sigma(2)(max) and the maximal adversarial variation Sigma(2)(max) max for strongly convex functions. Inspired by their work, we investigate the theoretical guarantees of optimistic online mirror descent (OMD) for the SEA model. For convex and smooth functions, we obtain the same O(root sigma(2)(1:T) + root Sigma(2)(1:T)) regret bound, without the convexity requirement of individual functions. For strongly convex and smooth functions, we establish an O(min{log(sigma(2)(1:T) + Sigma(2)(1:T)), (sigma(2)(max) + Sigma(2)(max)) log T}) bound, better than their O((sigma(2)(max) + Sigma(2)(max)) log T) result. For exp-concave and smooth functions, we achieve a new O(d log(sigma(2)(1:T) + Sigma(2)(1:T))) bound. Owing to the OMD framework, we further establish dynamic regret for convex and smooth functions, which is more favorable in non-stationary online scenarios.
引用
收藏
页数:34
相关论文
共 50 条
  • [1] Optimistic Online Mirror Descent for Bridging Stochastic and Adversarial Online Convex Optimization
    Chen, Sijia
    Zhang, Yu-Jie
    Tu, Wei-Wei
    Zhao, Peng
    Zhang, Lijun
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 62
  • [2] Distributed Mirror Descent for Online Composite Optimization
    Yuan, Deming
    Hong, Yiguang
    Ho, Daniel W. C.
    Xu, Shengyuan
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2021, 66 (02) : 714 - 729
  • [3] Faster Optimistic Online Mirror Descent for Extensive-Form Games
    Jiang, Huacong
    Liu, Weiming
    Li, Bin
    [J]. PRICAI 2022: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2022, 13629 : 90 - 103
  • [4] Online Convex Optimization with Stochastic Constraints
    Yu, Hao
    Neely, Michael J.
    Wei, Xiaohan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [5] Optimal distributed stochastic mirror descent for strongly convex optimization
    Yuan, Deming
    Hong, Yiguang
    Ho, Daniel W. C.
    Jiang, Guoping
    [J]. AUTOMATICA, 2018, 90 : 196 - 203
  • [6] Algorithms of Inertial Mirror Descent in Convex Problems of Stochastic Optimization
    Nazin, A. V.
    [J]. AUTOMATION AND REMOTE CONTROL, 2018, 79 (01) : 78 - 88
  • [7] Algorithms of Inertial Mirror Descent in Convex Problems of Stochastic Optimization
    A. V. Nazin
    [J]. Automation and Remote Control, 2018, 79 : 78 - 88
  • [8] Online convex optimization using coordinate descent algorithms
    Lin, Yankai
    Shames, Iman
    Nesic, Dragan
    [J]. AUTOMATICA, 2024, 165
  • [9] Between Stochastic and Adversarial Online Convex Optimization: Improved Regret Bounds via Smoothness
    Sachs, Sarah
    Hadiji, Hedi
    van Erven, Tim
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [10] Online Convex Optimization in Adversarial Markov Decision Processes
    Rosenberg, Aviv
    Mansour, Yishay
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97