ON MARGINAL SLICED INVERSE REGRESSION FOR ULTRAHIGH DIMENSIONAL MODEL-FREE FEATURE SELECTION

被引:20
|
作者
Yu, Zhou [1 ]
Dong, Yuexiao [2 ]
Shao, Jun [3 ]
机构
[1] East China Normal Univ, Sch Stat, Shanghai 200241, Peoples R China
[2] Temple Univ, Dept Stat, Philadelphia, PA 19122 USA
[3] Univ Wisconsin, Dept Stat, Madison, WI 53705 USA
来源
ANNALS OF STATISTICS | 2016年 / 44卷 / 06期
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Marginal coordinate test; sliced inverse regression; sufficient dimension reduction; sure independence screening; VARIABLE SELECTION; KOLMOGOROV FILTER; ADAPTIVE LASSO; REDUCTION; SHRINKAGE;
D O I
10.1214/15-AOS1424
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Model-free variable selection has been implemented under the sufficient dimension reduction framework since the seminal paper of Cook [Ann. Statist. 32 (2004) 1062-1092]. In this paper, we extend the marginal coordinate test for sliced inverse regression (SIR) in Cook (2004) and propose a novel marginal SIR utility for the purpose of ultrahigh dimensional feature selection. Two distinct procedures, Dantzig selector and sparse precision matrix estimation, are incorporated to get two versions of sample level marginal SIR utilities. Both procedures lead to model-free variable selection consistency with predictor dimensionality p diverging at an exponential rate of the sample size n. As a special case of marginal SIR, we ignore the correlation among the predictors and propose marginal independence SIR. Marginal independence SIR is closely related to many existing independence screening procedures in the literature, and achieves model-free screening consistency in the ultrahigh dimensional setting. The finite sample performances of the proposed procedures are studied through synthetic examples and an application to the small round blue cell tumors data.
引用
收藏
页码:2594 / 2623
页数:30
相关论文
共 50 条
  • [31] The cumulative Kolmogorov filter for model-free screening in ultrahigh dimensional data
    Kim, Arlene Kyoung Hee
    Shin, Seung Jun
    [J]. STATISTICS & PROBABILITY LETTERS, 2017, 126 : 238 - 243
  • [32] Variable importance assessment in sliced inverse regression for variable selection
    Jlassi, Ines
    Saracco, Jerome
    [J]. COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2019, 48 (01) : 169 - 199
  • [33] Sliced inverse regression for high-dimensional time series
    Becker, C
    Fried, R
    [J]. EXPLORATORY DATA ANALYSIS IN EMPIRICAL RESEARCH, PROCEEDINGS, 2003, : 3 - 11
  • [34] A NEW MODEL-FREE FEATURE SCREENING PROCEDURE FOR ULTRAHIGH-DIMENSIONAL INTERVAL-CENSORED FAILURE TIME DATA
    Zhang, Jing
    Du, Mingyue
    Liu, Yanyan
    Sun, Jianguo
    [J]. STATISTICA SINICA, 2023, 33 (03) : 1809 - 1830
  • [35] An algorithm with projection pursuit for sliced inverse regression model
    Mizuta, M
    Minami, H
    [J]. DATA ANALYSIS, CLASSIFICATION, AND RELATED METHODS, 2000, : 255 - 260
  • [36] Sparse sliced inverse regression for high dimensional data analysis
    Hilafu, Haileab
    Safo, Sandra E.
    [J]. BMC BIOINFORMATICS, 2022, 23 (01)
  • [37] Sparse sliced inverse regression for high dimensional data analysis
    Haileab Hilafu
    Sandra E. Safo
    [J]. BMC Bioinformatics, 23
  • [38] VALID POST-SELECTION INFERENCE IN MODEL-FREE LINEAR REGRESSION
    Kuchibhotla, Arun K.
    Brown, Lawrence D.
    Buja, Andreas
    Cai, Junhui
    George, Edward, I
    Zhao, Linda H.
    [J]. ANNALS OF STATISTICS, 2020, 48 (05): : 2953 - 2981
  • [39] MODEL-FREE COORDINATE TEST AND VARIABLE SELECTION VIA DIRECTIONAL REGRESSION
    Yu, Zhou
    Dong, Yuexiao
    [J]. STATISTICA SINICA, 2016, 26 (03) : 1159 - 1174
  • [40] An iterative model-free feature screening procedure: Forward recursive selection
    Xia, Siwei
    Yang, Yuehan
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 246