Contextual Inverse Optimization: Offline and Online Learning

被引:0
|
作者
Besbes, Omar [1 ]
Fonseca, Yuri [1 ]
Lobel, Ilan [2 ]
机构
[1] Columbia Univ, Grad Sch Business, Decis Risk & Operat, New York, NY 10027 USA
[2] NYU, Stern Sch Business, Technol Operat & Stat, New York, NY 10012 USA
关键词
contextual optimization; online optimization; imitation learning; inverse optimization; learning from revealed preferences; data-driven decision making;
D O I
10.1287/opre.2021.0369
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
We study the problems of offline and online contextual optimization with feedback information, where instead of observing the loss, we observe, after the fact, the optimal action an oracle with full knowledge of the objective function would have taken. We aim to minimize regret, which is defined as the difference between our losses and the ones incurred by an all-knowing oracle. In the offline setting, the decision maker has information available from past periods and needs to make one decision, whereas in the online setting, the decision maker optimizes decisions dynamically over time based a new set of feasible actions and contextual functions in each period. For the offline setting, we characterize the optimal minimax policy, establishing the performance that can be achieved as a function of the underlying geometry of the information induced by the data. In the online setting, we leverage this geometric characterization to optimize the cumulative regret. We develop an algorithm that yields the first regret bound for this problem that is logarithmic in the time horizon. Finally, we show via simulation that our proposed algorithms outperform previous methods from the literature.
引用
收藏
页数:21
相关论文
共 50 条
  • [1] Combined Online and Offline Inverse Dynamics Learning for a Robot Manipulator
    Panda, Amrut Sekhar
    Prakash, Ravi
    Behera, Laxmidhar
    Dutta, Ashish
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [2] Offline Contextual Bayesian Optimization
    Char, Ian
    Chung, Youngseog
    Neiswanger, Willie
    Kandasamy, Kirthevasan
    Nelson, Andrew Oakleigh
    Boyer, Mark D.
    Kolemen, Egemen
    Schneider, Jeff
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Community Exploration: From Offline Optimization to Online Learning
    Chen, Xiaowei
    Huang, Weiran
    Chen, Wei
    Lui, John C. S.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [4] Generalized Inverse Optimization through Online Learning
    Dong, Chaosheng
    Chen, Yiran
    Zeng, Bo
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] Online Learning versus Offline Learning
    Shai Ben-David
    Eyal Kushilevitz
    Yishay Mansour
    [J]. Machine Learning, 1997, 29 : 45 - 63
  • [6] Online learning versus offline learning
    BenDavid, S
    Kushilevitz, E
    Mansour, Y
    [J]. MACHINE LEARNING, 1997, 29 (01) : 45 - 63
  • [7] Hybrid Offline/Online Optimization for Energy Management via Reinforcement Learning
    Silvestri, Mattia
    De Filippo, Allegra
    Ruggeri, Federico
    Lombardi, Michele
    [J]. INTEGRATION OF CONSTRAINT PROGRAMMING, ARTIFICIAL INTELLIGENCE, AND OPERATIONS RESEARCH, CPAIOR 2022, 2022, 13292 : 358 - 373
  • [8] Emulating the Expert: Inverse Optimization through Online Learning
    Baermann, Andreas
    Pokutta, Sebastian
    Schneider, Oskar
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [9] Online versus offline learning
    Opper, M
    [J]. PHILOSOPHICAL MAGAZINE B-PHYSICS OF CONDENSED MATTER STATISTICAL MECHANICS ELECTRONIC OPTICAL AND MAGNETIC PROPERTIES, 1998, 77 (05): : 1531 - 1537
  • [10] Stateful Offline Contextual Policy Evaluation and Learning
    Kallus, Nathan
    Zhou, Angela
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151