ROBUST HIGH-DIMENSIONAL TUNING FREE MULTIPLE TESTING

被引:0
|
作者
Fan, Jianqing [1 ]
Lou, Zhipeng [2 ]
Yu, Mengxin [3 ]
机构
[1] Princeton Univ, Dept Operat Res & Financial Engn, Princeton, NJ 08544 USA
[2] Univ Pittsburgh, Dept Stat, Pittsburgh, PA USA
[3] Univ Penn, Dept Stat & Data Sci, Philadelphia, PA USA
来源
ANNALS OF STATISTICS | 2023年 / 51卷 / 05期
关键词
Key words and phrases. Robust statistical inference; heavy -tailed data; tuning free; weighted bootstrap; large; scale multiple testing; FALSE DISCOVERY RATE; RNA-SEQ; BOOTSTRAP APPROXIMATIONS; QUANTILE REGRESSION; EFFICIENT APPROACH; 2-SAMPLE TEST; U-STATISTICS; M-ESTIMATORS; REPRESENTATION; ASYMPTOTICS;
D O I
10.1214/23-AOS2322
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
A stylized feature of high-dimensional data is that many variables have heavy tails, and robust statistical inference is critical for valid large-scale statistical inference. Yet, the existing developments such as Winsorization, Huberization and median of means require the bounded second moments and involve variable-dependent tuning parameters, which hamper their fidelity in applications to large-scale problems. To liberate these constraints, this paper revisits the celebrated Hodges-Lehmann (HL) estimator for estimating location parameters in both the one- and two-sample problems, from a nonasymptotic perspective. Our study develops Berry-Esseen inequality and Cramer-type moderate deviation for the HL estimator based on newly developed nonasymptotic Bahadur representation and builds data-driven confidence intervals via a weighted bootstrap approach. These results allow us to extend the HL estimator to large-scale studies and propose tuning-free and moment-free high-dimensional inference procedures for testing global null and for large-scale multiple testing with false discovery proportion control. It is convincingly shown that the resulting tuning-free and moment-free methods control false discovery proportion at a prescribed level. The simulation studies lend further support to our developed theory.
引用
收藏
页码:2093 / 2115
页数:23
相关论文
共 50 条
  • [31] Improved Algorithms for High-dimensional Robust PCA
    Lin, Xiaoyong
    Zhang, Zeqiu
    Wang, Jue
    Zhang, Zhaoyang
    Qiu, Tingting
    Mi, Zhengkun
    2016 IEEE INTERNATIONAL CONFERENCE ON UBIQUITOUS WIRELESS BROADBAND (ICUWB2016), 2016,
  • [32] A robust PCR method for high-dimensional regressors
    Hubert, M
    Verboven, S
    JOURNAL OF CHEMOMETRICS, 2003, 17 (8-9) : 438 - 452
  • [33] Robust Methods for High-Dimensional Linear Learning
    Merad, Ibrahim
    Gaiffas, Stephane
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [34] Fast Robust Correlation for High-Dimensional Data
    Raymaekers, Jakob
    Rousseeuw, Peter J.
    TECHNOMETRICS, 2021, 63 (02) : 184 - 198
  • [35] Is high-dimensional photonic entanglement robust to noise?
    Zhu, F.
    Tyler, M.
    Valencia, N. H.
    Malik, M.
    Leach, J.
    AVS QUANTUM SCIENCE, 2021, 3 (01):
  • [36] Scale calibration for high-dimensional robust regression
    Loh, Po-Ling
    ELECTRONIC JOURNAL OF STATISTICS, 2021, 15 (02): : 5933 - 5994
  • [37] Robust Ridge Regression for High-Dimensional Data
    Maronna, Ricardo A.
    TECHNOMETRICS, 2011, 53 (01) : 44 - 53
  • [38] Superposed Atomic Representation for Robust High-Dimensional Data Recovery of Multiple Low-Dimensional Structures
    Wang, Yulong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 15735 - 15742
  • [39] A study on tuning parameter selection for the high-dimensional lasso
    Homrighausen, Darren
    McDonald, Daniel J.
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2018, 88 (15) : 2865 - 2892
  • [40] A Survey of Tuning Parameter Selection for High-Dimensional Regression
    Wu, Yunan
    Wang, Lan
    ANNUAL REVIEW OF STATISTICS AND ITS APPLICATION, VOL 7, 2020, 2020, 7 : 209 - 226