Targeted L1-Regularization and Joint Modeling of Neural Networks for Causal Inference

被引:1
|
作者
Rostami, Mehdi [1 ]
Saarela, Olli [1 ]
机构
[1] Univ Toronto, Dalla Lana Sch Publ Hlth, Toronto, ON M5T 3M7, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
causal Inference; instrumental variables; neural networks; doubly robust estimation; regularization;
D O I
10.3390/e24091290
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
The calculation of the Augmented Inverse Probability Weighting (AIPW) estimator of the Average Treatment Effect (ATE) is carried out in two steps, where in the first step, the treatment and outcome are modeled, and in the second step, the predictions are inserted into the AIPW estimator. The model misspecification in the first step has led researchers to utilize Machine Learning algorithms instead of parametric algorithms. However, the existence of strong confounders and/or Instrumental Variables (IVs) can lead the complex ML algorithms to provide perfect predictions for the treatment model which can violate the positivity assumption and elevate the variance of AIPW estimators. Thus the complexity of ML algorithms must be controlled to avoid perfect predictions for the treatment model while still learning the relationship between the confounders and the treatment and outcome. We use two NN architectures with an L-1-regularization on specific NN parameters and investigate how their certain hyperparameters should be tuned in the presence of confounders and Ws to achieve a low bias-variance tradeoff for ATE estimators such as AIPW estimator. Through simulation results, we will provide recommendations as to how NNs can be employed for ATE estimation.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] Estimating Sparse Volterra Models using Group L1-Regularization
    Song, Dong
    Wang, Haonan
    Berger, Theodore W.
    2010 ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2010, : 4128 - 4131
  • [22] Sparse Analog Associative Memory via L1-Regularization and Thresholding
    Chalasani, Rakesh
    Principe, Jose C.
    2011 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2011, : 1994 - 1999
  • [23] High-Dimensional Variable Selection With Reciprocal L1-Regularization
    Song, Qifan
    Liang, Faming
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2015, 110 (512) : 1607 - 1620
  • [24] CONVERGENCE RATES FOR l1-REGULARIZATION WITHOUT THE HELP OF A VARIATIONAL INEQUALITY
    Gerth, Daniel
    ELECTRONIC TRANSACTIONS ON NUMERICAL ANALYSIS, 2017, 46 : 233 - 244
  • [25] Formal convergence analysis on deterministic l1-regularization based mini-batch learning for RBF networks
    Liu, Zhaofeng
    Leung, Chi-Sing
    So, Hing Cheung
    NEUROCOMPUTING, 2023, 532 : 77 - 93
  • [26] On Accelerated Gradient Approximation for Least Square Regression with L1-regularization
    Zhang, Yongquan
    Sun, Jianyong
    2015 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2015, : 1569 - 1575
  • [27] On l1-Regularization Under Continuity of the Forward Operator in Weaker Topologies
    Gerth, Daniel
    Hofmann, Bernd
    NEW TRENDS IN PARAMETER IDENTIFICATION FOR MATHEMATICAL MODELS, 2018, : 67 - 88
  • [28] Conditional Quantile Regression with l1-regularization and ε-insensitive Pinball Loss
    Li, Meng
    Zhang, Meijian
    Sun, Hongwei
    2015 8TH INTERNATIONAL CONFERENCE ON BIOMEDICAL ENGINEERING AND INFORMATICS (BMEI), 2015, : 843 - 851
  • [29] Estimation of parameters and delay in driver models using L1-regularization
    Hosseini, SeyedMehrdad
    Koroglu, Hakan
    Sjoberg, Jonas
    2016 EUROPEAN CONTROL CONFERENCE (ECC), 2016, : 945 - 950
  • [30] On l1-Regularization in Light of Nashed's Ill-Posedness Concept
    Flemming, Jens
    Hofmann, Bernd
    Veselic, Ivan
    COMPUTATIONAL METHODS IN APPLIED MATHEMATICS, 2015, 15 (03) : 279 - 289