Regret Minimization in Stochastic Non-Convex Learning via a Proximal-Gradient Approach

被引:0
|
作者
Hallak, Nadav [1 ]
Mertikopoulos, Panayotis [2 ,3 ]
Cevher, Volkan [4 ]
机构
[1] Technion, Fac Ind Engn & Management, Haifa, Israel
[2] Univ Grenoble Alpes, CNRS, INRIA, LIG, Grenoble, France
[3] Criteo AI Lab, Grenoble, France
[4] Ecole Polytech Fed Lausanne EPFL, Lausanne, Switzerland
基金
瑞士国家科学基金会; 欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper develops a methodology for regret minimization with stochastic first-order oracle feedback in online, constrained, non-smooth, non-convex problems. In this setting, the minimization of external regret is beyond reach for first-order methods, and there are no gradient-based algorithmic frameworks capable of providing a solution. On that account, we focus on a local regret measure defined via a proximal-gradient mapping, that also encompasses the original notion proposed by Hazan et al, (2017). To achieve no local regret in this setting, we develop a proximal-gradient method based on stochastic first-order feedback, and a simpler method for when access to a perfect first-order oracle is possible. Both methods are order-optimal (in the min-max sense), and we also establish a bound on the number of proximal-gradient queries these methods require. As an important application of our results, we also obtain a link between online and offline non-convex stochastic optimization manifested as a new proximal-gradient scheme with complexity guarantees matching those obtained via variance reduction techniques.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Stochastic variance reduced gradient with hyper-gradient for non-convex large-scale learning
    Zhuang Yang
    Applied Intelligence, 2023, 53 : 28627 - 28641
  • [32] A proximal iterative approach to a non-convex optimization problem
    Moudafi, A.
    NONLINEAR ANALYSIS-THEORY METHODS & APPLICATIONS, 2010, 72 (02) : 704 - 709
  • [33] Scaling up stochastic gradient descent for non-convex optimisation
    Mohamad, Saad
    Alamri, Hamad
    Bouchachia, Abdelhamid
    MACHINE LEARNING, 2022, 111 (11) : 4039 - 4079
  • [34] Scaling up stochastic gradient descent for non-convex optimisation
    Saad Mohamad
    Hamad Alamri
    Abdelhamid Bouchachia
    Machine Learning, 2022, 111 : 4039 - 4079
  • [35] A splitting bundle approach for non-smooth non-convex minimization
    Fuduli, A.
    Gaudioso, M.
    Nurminski, E. A.
    OPTIMIZATION, 2015, 64 (05) : 1131 - 1151
  • [36] Inexact Proximal Gradient Methods for Non-Convex and Non-Smooth Optimization
    Gu, Bin
    Wang, De
    Huo, Zhouyuan
    Huang, Heng
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3093 - 3100
  • [37] DISTRIBUTED PROXIMAL-GRADIENT METHOD FOR CONVEX OPTIMIZATION WITH INEQUALITY CONSTRAINTS
    Li, Jueyou
    Wu, Changzhi
    Wu, Zhiyou
    Long, Qiang
    Wang, Xiangyu
    ANZIAM JOURNAL, 2014, 56 (02): : 160 - 178
  • [38] Non-Convex Feature Learning via Operator
    Kong, Deguang
    Ding, Chris
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 1918 - 1924
  • [39] Implementation of reduced gradient with bisection algorithms for non-convex optimization problem via stochastic perturbation
    Abdelkrim El Mouatasim
    Numerical Algorithms, 2018, 78 : 41 - 62