Online Learning With Inexact Proximal Online Gradient Descent Algorithms

被引:59
|
作者
Dixit, Rishabh [1 ]
Bedi, Unlit Singh [1 ]
Tripathi, Ruchi [1 ]
Rajawat, Ketan [1 ]
机构
[1] Indian Inst Technol Kanpur, Dept Elect Engn, Kanpur 208016, Uttar Pradesh, India
关键词
Dynamic regret; gradient descent; online convex optimization; subspace tracking; SUBGRADIENT METHODS; STOCHASTIC METHODS; ROBUST PCA;
D O I
10.1109/TSP.2018.2890368
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We consider nondifferentiable dynamic optimization problems such as those arising in robotics and subspace tracking. Given the computational constraints and the time-varying nature of the problem, a low-complexity algorithm is desirable, while the accuracy of the solution may only increase slowly over time. We put forth the proximal online gradient descent (OGD) algorithm for tracking the optimum of a composite objective function comprising of a differentiable loss function and a nondifferentiable regularizer. An online learning framework is considered and the gradient of the loss function is allowed to be erroneous. Both, the gradient error as well as the dynamics of the function optimum or target are adversarial and the performance of the inexact proximal OGD is characterized in terms of its dynamic regret, expressed in terms of the cumulative error and path length of the target. The proposed inexact proximal OGD is generalized for application to large-scale problems where the loss function has a finite sum structure. In such cases, evaluation of the full gradient may not be viable and a variance reduced version is proposed that allows the component functions to be subsampled. The efficacy of the proposed algorithms is tested on the problem of formation control in robotics and on the dynamic foreground-background separation problem in video.
引用
收藏
页码:1338 / 1352
页数:15
相关论文
共 50 条
  • [41] Online Performative Gradient Descent for Learning Nash Equilibria in Decision-Dependent Games
    Zhu, Zihan
    Fang, Ethan X.
    Yang, Zhuoran
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [42] Online convex optimization in the bandit setting: gradient descent without a gradient
    Flaxman, Abraham D.
    Kalai, Adam Tauman
    McMahan, H. Brendan
    [J]. PROCEEDINGS OF THE SIXTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, 2005, : 385 - 394
  • [43] Online Learning Over Dynamic Graphs via Distributed Proximal Gradient Algorithm
    Dixit, Rishabh
    Bedi, Amrit Singh
    Rajawat, Ketan
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2021, 66 (11) : 5065 - 5079
  • [44] Online convex optimization using coordinate descent algorithms
    Lin, Yankai
    Shames, Iman
    Nesic, Dragan
    [J]. AUTOMATICA, 2024, 165
  • [45] Online Pairwise Learning Algorithms
    Ying, Yiming
    Zhou, Ding-Xuan
    [J]. NEURAL COMPUTATION, 2016, 28 (04) : 743 - 777
  • [46] Efficient inexact proximal gradient algorithms for structured sparsity-inducing norm
    Gu, Bin
    Geng, Xiang
    Li, Xiang
    Zheng, Guansheng
    [J]. NEURAL NETWORKS, 2019, 118 : 352 - 362
  • [47] A basic formula for Online policy gradient algorithms
    Cao, XR
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2005, 50 (05) : 696 - 699
  • [48] ONLINE REGULARIZED GENERALIZED GRADIENT CLASSIFICATION ALGORITHMS
    Leilei Zhang (Ningbo University
    [J]. Analysis in Theory and Applications, 2010, 26 (03) : 278 - 300
  • [49] Regret Analysis of Online Gradient Descent-based Iterative Learning Control with Model Mismatch
    Balta, Efe C.
    Iannelli, Andrea
    Smith, Roy S.
    Lygeros, John
    [J]. 2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 1479 - 1484
  • [50] The Inexact Cyclic Block Proximal Gradient Method and Properties of Inexact Proximal Maps
    Leandro Farias Maia
    David Huckleberry Gutman
    Ryan Christopher Hughes
    [J]. Journal of Optimization Theory and Applications, 2024, 201 : 668 - 698