Generalization Bounds in the Predict-Then-Optimize Framework

被引:0
|
作者
El Balghiti, Othman [1 ]
Elmachtoub, Adam N. [1 ]
Grigas, Paul [2 ]
Tewari, Ambuj [3 ]
机构
[1] Columbia Univ, Dept Ind Engn & Operat Res, New York, NY 10027 USA
[2] Univ Calif Berkeley, Dept Ind Engn & Operat Res, Berkeley, CA 94720 USA
[3] Univ Michigan, Dept Stat, Ann Arbor, MI 48109 USA
基金
美国国家科学基金会;
关键词
generalization bounds; prescriptive analytics; regression; predict-then-optimize; GENERALIZATION ERROR; SMART PREDICT;
D O I
10.1287/moor.2022.1330
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
The predict-then-optimize framework is fundamental in many practical settings: predict the unknown parameters of an optimization problem and then solve the problem using the predicted values of the parameters. A natural loss function in this environment is to consider the cost of the decisions induced by the predicted parameters in contrast to the prediction error of the parameters. This loss function is referred to as the smart predict then-optimize (SPO) loss. In this work, we seek to provide bounds on how well the performance of a prediction model fit on training data generalizes out of sample in the context of the SPO loss. Because the SPO loss is nonconvex and non-Lipschitz, standard results for deriving generalization bounds do not apply. We first derive bounds based on the Natarajan dimension that, in the case of a polyhedral feasible region, scale at most logarithmically in the number of extreme points but, in the case of a general convex feasible region, have linear dependence on the decision dimension. By exploiting the structure of the SPO loss function and a key property of the feasible region, which we denote as the strength property, we can dramatically improve the dependence on the decision and feature dimensions. Our approach and analysis rely on placing a margin around problematic predictions that do not yield unique optimal solutions and then providing generalization bounds in the context of a modified margin SPO loss function that is Lipschitz continuous. Finally, we characterize the strength property and show that the modified SPO loss can be computed efficiently for both strongly convex bodies and polytopes with an explicit extreme point representation.
引用
下载
收藏
页码:2043 / 2065
页数:24
相关论文
共 50 条
  • [21] Learning to Predict and Optimize Imperfect MIMO System Performance: Framework and Application
    Su, Jingyi
    Meng, Fan
    Liu, Shengheng
    Huang, Yongming
    Lu, Zhaohua
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 335 - 340
  • [22] On Generalization Bounds for Projective Clustering
    Bucarelli, Maria Sofia
    Larsen, Matilde Fjeldsø
    Schwiegelshohn, Chris
    Toftrup, Mads Bech
    Advances in Neural Information Processing Systems, 2023, 36
  • [23] New bounds for correct generalization
    Mattera, D
    Palmieri, F
    1997 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, 1997, : 1051 - 1055
  • [24] Generalization bounds for averaged classifiers
    Freund, Y
    Mansour, Y
    Schapire, RE
    ANNALS OF STATISTICS, 2004, 32 (04): : 1698 - 1722
  • [25] Comparing Comparators in Generalization Bounds
    Hellstrom, Fredrik
    Guedj, Benjamin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [26] Generalization bounds of incremental SVM
    Zeng, Jingjing
    Zou, Bin
    Qin, Yimo
    Xu, Jie
    INTERNATIONAL JOURNAL OF WAVELETS MULTIRESOLUTION AND INFORMATION PROCESSING, 2024,
  • [27] Margin distribution bounds on generalization
    Shawe-Taylor, J
    Cristianini, N
    COMPUTATIONAL LEARNING THEORY, 1999, 1572 : 263 - 273
  • [28] Sharper Generalization Bounds for Clustering
    Li, Shaojie
    Liu, Yong
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [29] Geometric bounds for generalization in boosting
    Mannor, S
    Meir, R
    COMPUTATIONAL LEARNING THEORY, PROCEEDINGS, 2001, 2111 : 461 - 472
  • [30] Information Complexity and Generalization Bounds
    Banerjee, Pradeep Kr
    Montufar, Guido
    2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 676 - 681