Minimax-Optimal Policy Learning Under Unobserved Confounding

被引:0
|
作者
Kallus, Nathan [1 ]
Zhou, Angela [1 ]
机构
[1] Cornell Univ, New York, NY 10044 USA
基金
美国国家科学基金会;
关键词
policy learning; optimization; causal inference; personalized medicine; data-driven decision making; REGRET TREATMENT CHOICE; SENSITIVITY-ANALYSIS; IDENTIFICATION; THERAPY; TRIALS; ISSUES;
D O I
10.1287/mnsc.2020.3699
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
We study the problem of learning personalized decision policies from observational data while accounting for possible unobserved confounding. Previous approaches, which assume unconfoundedness, that is, that no unobserved confounders affect both the treatment assignment as well as outcome, can lead to policies that introduce harm rather than benefit when some unobserved confounding is present as is generally the case with observational data. Instead, because policy value and regret may not be point-identifiable, we study a method that minimizes the worst-case estimated regret of a candidate policy against a baseline policy over an uncertainty set for propensity weights that controls the extent of unobserved confounding. We prove generalization guarantees that ensure our policy is safe when applied in practice and in fact obtains the best possible uniform control on the range of all possible population regrets that agree with the possible extent of confounding. We develop efficient algorithmic solutions to compute this minimax-optimal policy. Finally, we assess and compare our methods on synthetic and semisynthetic data. In particular, we consider a case study on personalizing hormone replacement therapy based on observational data, in which we validate our results on a randomized experiment. We demonstrate that hidden confounding can hinder existing policy-learning approaches and lead to unwarranted harm although our robust approach guarantees safety and focuses on well-evidenced improvement, a necessity for making personalized treatment policies learned from observational data reliable in practice.
引用
收藏
页码:2870 / 2890
页数:21
相关论文
共 50 条
  • [1] Learning Minimax-Optimal Terminal State Estimators and Smoothers
    Zhang, Xiangyuan
    Velicheti, Raj Kiriti
    Basar, Tamer
    [J]. IFAC PAPERSONLINE, 2023, 56 (02): : 11545 - 11550
  • [2] Minimax-Optimal Location Estimation
    Gupta, Shivam
    Lee, Jasper C. H.
    Price, Eric
    Valiant, Paul
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Minimax-Optimal Off-Policy Evaluation with Linear Function Approximation
    Duan, Yaqi
    Jia, Zeyu
    Wang, Mengdi
    [J]. 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [4] Minimax-Optimal Off-Policy Evaluation with Linear Function Approximation
    Duan, Yaqi
    Jia, Zeyu
    Wang, Mengdi
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [5] Analysis and design of minimax-optimal interpolators
    Choi, H
    Munson, DC
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1998, 46 (06) : 1571 - 1579
  • [6] Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding
    Namkoong, Hongseok
    Keramati, Ramtin
    Yadlowsky, Steve
    Brunskill, Emma
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [7] Minimax-optimal classification with dyadic decision trees
    Scott, C
    Nowak, RD
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2006, 52 (04) : 1335 - 1353
  • [8] Minimax-optimal Inference from Partial Rankings
    Hajek, Bruce
    Oh, Sewoong
    Xu, Jiaming
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [9] Nearly Minimax-Optimal Regret for Linearly Parameterized Bandits
    Li, Yingkai
    Wang, Yining
    Zhou, Yuan
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2024, 70 (01) : 372 - 388
  • [10] MINIMAX-OPTIMAL NONPARAMETRIC REGRESSION IN HIGH DIMENSIONS
    Yang, Yun
    Tokdar, Surya T.
    [J]. ANNALS OF STATISTICS, 2015, 43 (02): : 652 - 674