No-Regret Online Prediction with Strategic Experts

被引:0
|
作者
Sadeghi, Omid [1 ]
Fazel, Maryam [1 ]
机构
[1] Univ Washington, Seattle, WA 98195 USA
关键词
BOUNDS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study a generalization of the online binary prediction with expert advice framework where at each round, the learner is allowed to pick m >= 1 experts from a pool of K experts and the overall utility is a modular or submodular function of the chosen experts. We focus on the setting in which experts act strategically and aim to maximize their influence on the algorithm's predictions by potentially misreporting their beliefs about the events. Among others, this setting finds applications in forecasting competitions where the learner seeks not only to make predictions by aggregating different forecasters but also to rank them according to their relative performance. Our goal is to design algorithms that satisfy the following two requirements: 1) Incentive-compatible: Incentivize the experts to report their beliefs truthfully, and 2) No-regret: Achieve sublinear regret with respect to the true beliefs of the best-fixed set of m experts in hindsight. Prior works have studied this framework when m = 1 and provided incentive-compatible no-regret algorithms for the problem. We first show that a simple reduction of our problem to the m = 1 setting is neither efficient nor effective. Then, we provide algorithms that utilize the specific structure of the utility functions to achieve the two desired goals.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] No-Regret Caching via Online Mirror Descent
    Salem, Tareq Si
    Neglia, Giovanni
    Ioannidis, Straus
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [2] No-Regret and Incentive-Compatible Online Learning
    Freeman, Rupert
    Pennock, David M.
    Podimata, Chara
    Vaughan, Jennifer Wortman
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [3] No-regret Online Learning over Riemannian Manifolds
    Wang, Xi
    Tu, Zhipeng
    Hong, Yiguang
    Wu, Yingyi
    Shi, Guodong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [4] No-Regret and Incentive-Compatible Online Learning
    Freeman, Rupert
    Pennock, David M.
    Podimata, Chara
    Vaughan, Jennifer Wortman
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [5] No-regret Caching via Online Mirror Descent
    Salem, Tareq Si
    Neglia, Giovanni
    Ioannidis, Stratis
    ACM TRANSACTIONS ON MODELING AND PERFORMANCE EVALUATION OF COMPUTING SYSTEMS, 2023, 8 (04)
  • [6] No-regret algorithms for online k-submodular maximization
    Soma, Tasuku
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [7] A Reduction from Reinforcement Learning to No-Regret Online Learning
    Cheng, Ching-An
    des Combes, Remi Tachet
    Boots, Byron
    Gordon, Geoff
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 3514 - 3523
  • [8] No-Regret Online Reinforcement Learning with Adversarial Losses and Transitions
    Jin, Tiancheng
    Liu, Junyan
    Rouyer, Chloe
    Chang, William
    Wei, Chen-Yu
    Luo, Haipeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] No-regret boosting
    Gambin, Anna
    Szczurek, Ewa
    ADAPTIVE AND NATURAL COMPUTING ALGORITHMS, PT 1, 2007, 4431 : 422 - +
  • [10] NO-REGRET NON-CONVEX ONLINE META-LEARNING
    Zhuang, Zhenxun
    Wang, Yunlong
    Yu, Kezi
    Lu, Songtao
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3942 - 3946