No-regret learning in games with noisy feedback: Faster rates and adaptivity via learning rate separation

被引:0
|
作者
Hsieh, Yu-Guan [1 ]
Antonakopoulos, Kimon [2 ]
Cevher, Volkan [2 ]
Mertikopoulos, Panayotis [1 ,3 ,4 ]
机构
[1] Univ Grenoble Alpes, Grenoble, France
[2] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[3] CNRS, Inria, LIG, Paris, France
[4] Criteo AI Lab, Ann Arbor, MI USA
基金
瑞士国家科学基金会; 欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We examine the problem of regret minimization when the learner is involved in a continuous game with other optimizing agents: in this case, if all players follow a no-regret algorithm, it is possible to achieve significantly lower regret relative to fully adversarial environments. We study this problem in the context of variationally stable games (a class of continuous games which includes all convex-concave and monotone games), and when the players only have access to noisy estimates of their individual payoff gradients. If the noise is additive, the game-theoretic and purely adversarial settings enjoy similar regret guarantees; however, if the noise is multiplicative, we show that the learners can, in fact, achieve constant regret. We achieve this faster rate via an optimistic gradient scheme with learning rate separation - that is, the method's extrapolation and update steps are tuned to different schedules, depending on the noise profile. Subsequently, to eliminate the need for delicate hyperparameter tuning, we propose a fully adaptive method that attains nearly the same guarantees as its non-adapted counterpart, while operating without knowledge of either the game or of the noise profile.
引用
收藏
页数:13
相关论文
共 37 条
  • [1] No-Regret Learning in Bayesian Games
    Hartline, Jason
    Syrgkanis, Vasilis
    Tardos, Eva
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [2] Distributed No-Regret Learning in Aggregative Games With Residual Bandit Feedback
    Liu, Wenting
    Yi, Peng
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2024, 11 (04): : 1734 - 1745
  • [3] Limits and limitations of no-regret learning in games
    Monnot, Barnabe
    Piliouras, Georgios
    KNOWLEDGE ENGINEERING REVIEW, 2017, 32
  • [4] No-Regret Learning in Dynamic Stackelberg Games
    Lauffer, Niklas
    Ghasemi, Mahsa
    Hashemi, Abolfazl
    Savas, Yagiz
    Topcu, Ufuk
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (03) : 1418 - 1431
  • [5] Doubly Optimal No-Regret Learning in Monotone Games
    Cai, Yang
    Zheng, Weiqiang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [6] No-Regret Learning and Equilibrium Computation in Quantum Games
    Lin, Wayne
    Piliouras, Georgios
    Sim, Ryann
    Varvitsiotis, Antonios
    QUANTUM, 2024, 8
  • [7] Doubly Optimal No-Regret Online Learning in Strongly Monotone Games with Bandit Feedback
    Ba, Wenjia
    Lin, Tianyi
    Zhang, Jiawei
    Zhou, Zhengyuan
    OPERATIONS RESEARCH, 2025,
  • [8] No-Regret Learning in Unknown Games with Correlated Payoffs
    Sessa, Pier Giuseppe
    Bogunovic, Ilija
    Kamgarpour, Maryam
    Krause, Andreas
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Weighted Voting Via No-Regret Learning
    Haghtalab, Nika
    Noothigattu, Ritesh
    Procaccia, Ariel D.
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1055 - 1062
  • [10] Sampling Equilibria: Fast No-Regret Learning in Structured Games
    Beaglehole, Daniel
    Hopkins, Max
    Kane, Daniel
    Liu, Sihan
    Lovett, Shachar
    PROCEEDINGS OF THE 2023 ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, SODA, 2023, : 3817 - 3855