Regret Minimization in Stochastic Non-Convex Learning via a Proximal-Gradient Approach

被引:0
|
作者
Hallak, Nadav [1 ]
Mertikopoulos, Panayotis [2 ,3 ]
Cevher, Volkan [4 ]
机构
[1] Technion, Fac Ind Engn & Management, Haifa, Israel
[2] Univ Grenoble Alpes, CNRS, INRIA, LIG, Grenoble, France
[3] Criteo AI Lab, Grenoble, France
[4] Ecole Polytech Fed Lausanne EPFL, Lausanne, Switzerland
基金
瑞士国家科学基金会; 欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper develops a methodology for regret minimization with stochastic first-order oracle feedback in online, constrained, non-smooth, non-convex problems. In this setting, the minimization of external regret is beyond reach for first-order methods, and there are no gradient-based algorithmic frameworks capable of providing a solution. On that account, we focus on a local regret measure defined via a proximal-gradient mapping, that also encompasses the original notion proposed by Hazan et al, (2017). To achieve no local regret in this setting, we develop a proximal-gradient method based on stochastic first-order feedback, and a simpler method for when access to a perfect first-order oracle is possible. Both methods are order-optimal (in the min-max sense), and we also establish a bound on the number of proximal-gradient queries these methods require. As an important application of our results, we also obtain a link between online and offline non-convex stochastic optimization manifested as a new proximal-gradient scheme with complexity guarantees matching those obtained via variance reduction techniques.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] NON-CONVEX MINIMIZATION PROBLEMS
    EKELAND, I
    BULLETIN OF THE AMERICAN MATHEMATICAL SOCIETY, 1979, 1 (03) : 443 - 474
  • [22] Stochastic proximal-gradient algorithms for penalized mixed models
    Gersende Fort
    Edouard Ollier
    Adeline Samson
    Statistics and Computing, 2019, 29 : 231 - 253
  • [23] Stochastic proximal-gradient algorithms for penalized mixed models
    Fort, Gersende
    Ollier, Edouard
    Samson, Adeline
    STATISTICS AND COMPUTING, 2019, 29 (02) : 231 - 253
  • [24] Provable Efficient Online Matrix Completion via Non-convex Stochastic Gradient Descent
    Jin, Chi
    Kakade, Sham M.
    Netrapalli, Praneeth
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [25] A Distributed Stochastic Proximal-Gradient Algorithm for Composite Optimization
    Niu, Youcheng
    Li, Huaqing
    Wang, Zheng
    Lu, Qingguo
    Xia, Dawen
    Ji, Lianghao
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2021, 8 (03): : 1383 - 1393
  • [26] Stochastic proximal linear method for structured non-convex problems
    Hazan, Tamir
    Sabach, Shoham
    Voldman, Sergey
    OPTIMIZATION METHODS & SOFTWARE, 2020, 35 (05): : 921 - 937
  • [27] Taming Convergence for Asynchronous Stochastic Gradient Descent with Unbounded Delay in Non-Convex Learning
    Zhang, Xin
    Liu, Jia
    Zhu, Zhengyuan
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 3580 - 3585
  • [28] A stochastic recursive gradient algorithm with inertial extrapolation for non-convex problems and machine learning
    Mo, Zhongyu
    Ouyang, Chen
    Pham, Hongtruong
    Yuan, Gonglin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025,
  • [29] Stochastic variance reduced gradient with hyper-gradient for non-convex large-scale learning
    Yang, Zhuang
    APPLIED INTELLIGENCE, 2023, 53 (23) : 28627 - 28641
  • [30] Fast Proximal Gradient Descent for A Class of Non-convex and Non-smooth Sparse Learning Problems
    Yang, Yingzhen
    Yu, Jiahui
    35TH UNCERTAINTY IN ARTIFICIAL INTELLIGENCE CONFERENCE (UAI 2019), 2020, 115 : 1253 - 1262