Generalization Bounds in the Predict-Then-Optimize Framework

被引:0
|
作者
El Balghiti, Othman [1 ]
Elmachtoub, Adam N. [1 ]
Grigas, Paul [2 ]
Tewari, Ambuj [3 ]
机构
[1] Columbia Univ, Dept Ind Engn & Operat Res, New York, NY 10027 USA
[2] Univ Calif Berkeley, Dept Ind Engn & Operat Res, Berkeley, CA 94720 USA
[3] Univ Michigan, Dept Stat, Ann Arbor, MI 48109 USA
基金
美国国家科学基金会;
关键词
generalization bounds; prescriptive analytics; regression; predict-then-optimize; GENERALIZATION ERROR; SMART PREDICT;
D O I
10.1287/moor.2022.1330
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
The predict-then-optimize framework is fundamental in many practical settings: predict the unknown parameters of an optimization problem and then solve the problem using the predicted values of the parameters. A natural loss function in this environment is to consider the cost of the decisions induced by the predicted parameters in contrast to the prediction error of the parameters. This loss function is referred to as the smart predict then-optimize (SPO) loss. In this work, we seek to provide bounds on how well the performance of a prediction model fit on training data generalizes out of sample in the context of the SPO loss. Because the SPO loss is nonconvex and non-Lipschitz, standard results for deriving generalization bounds do not apply. We first derive bounds based on the Natarajan dimension that, in the case of a polyhedral feasible region, scale at most logarithmically in the number of extreme points but, in the case of a general convex feasible region, have linear dependence on the decision dimension. By exploiting the structure of the SPO loss function and a key property of the feasible region, which we denote as the strength property, we can dramatically improve the dependence on the decision and feature dimensions. Our approach and analysis rely on placing a margin around problematic predictions that do not yield unique optimal solutions and then providing generalization bounds in the context of a modified margin SPO loss function that is Lipschitz continuous. Finally, we characterize the strength property and show that the modified SPO loss can be computed efficiently for both strongly convex bodies and polytopes with an explicit extreme point representation.
引用
下载
收藏
页码:2043 / 2065
页数:24
相关论文
共 50 条
  • [1] Generalization Bounds in the Predict-then-Optimize Framework
    El Balghiti, Othman
    Elmachtoub, Adam N.
    Grigas, Paul
    Tewari, Ambuj
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Risk Bounds and Calibration for a Smart Predict-then-Optimize Method
    Liu, Heyuan
    Grigas, Paul
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [3] Decision Trees for Decision-Making under the Predict-then-Optimize Framework
    Elmachtoub, Adam N.
    Liang, Jason Cheuk Nam
    McNellis, Ryan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [4] Decision Trees for Decision-Making under the Predict-then-Optimize Framework
    Elmachtoub, Adam N.
    Liang, Jason Cheuk Nam
    McNellis, Ryan
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [5] Data Transformation in the Predict-Then-Optimize Framework: Enhancing Decision Making under Uncertainty
    Tian, Xuecheng
    Guan, Yanxia
    Wang, Shuaian
    MATHEMATICS, 2023, 11 (17)
  • [6] A Predict-Then-Optimize Couriers Allocation Framework for Emergency Last-mile Logistics
    Xia, Kaiwen
    Lin, Li
    Wang, Shuai
    Wang, Haotian
    Zhang, Desheng
    He, Tian
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5237 - 5248
  • [7] Characterizing and Improving the Robustness of Predict-Then-Optimize Frameworks
    Johnson-Yu, Sonja
    Finocchiaro, Jessie
    Wang, Kai
    Vorobeychik, Yevgeniy
    Sinha, Arunesh
    Taneja, Aparna
    Tambe, Milind
    DECISION AND GAME THEORY FOR SECURITY, GAMESEC 2023, 2023, 14167 : 133 - 152
  • [8] A Deficiency of the Predict-Then-Optimize Framework: Decreased Decision Quality with Increased Data Size
    Wang, Shuaian
    Tian, Xuecheng
    MATHEMATICS, 2023, 11 (15)
  • [9] Predict-then-optimize or predict-and-optimize? An empirical evaluation of cost-sensitive learning strategies
    Vanderschueren, Toon
    Verdonck, Tim
    Baesens, Bart
    Verbeke, Wouter
    INFORMATION SCIENCES, 2022, 594 : 400 - 415
  • [10] Leaving the Nest: Going beyond Local Loss Functions for Predict-Then-Optimize
    Shah, Sanket
    Wilder, Bryan
    Perrault, Andrew
    Tambe, Milind
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14902 - 14909