ROBUST HIGH-DIMENSIONAL TUNING FREE MULTIPLE TESTING

被引:0
|
作者
Fan, Jianqing [1 ]
Lou, Zhipeng [2 ]
Yu, Mengxin [3 ]
机构
[1] Princeton Univ, Dept Operat Res & Financial Engn, Princeton, NJ 08544 USA
[2] Univ Pittsburgh, Dept Stat, Pittsburgh, PA USA
[3] Univ Penn, Dept Stat & Data Sci, Philadelphia, PA USA
来源
ANNALS OF STATISTICS | 2023年 / 51卷 / 05期
关键词
Key words and phrases. Robust statistical inference; heavy -tailed data; tuning free; weighted bootstrap; large; scale multiple testing; FALSE DISCOVERY RATE; RNA-SEQ; BOOTSTRAP APPROXIMATIONS; QUANTILE REGRESSION; EFFICIENT APPROACH; 2-SAMPLE TEST; U-STATISTICS; M-ESTIMATORS; REPRESENTATION; ASYMPTOTICS;
D O I
10.1214/23-AOS2322
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
A stylized feature of high-dimensional data is that many variables have heavy tails, and robust statistical inference is critical for valid large-scale statistical inference. Yet, the existing developments such as Winsorization, Huberization and median of means require the bounded second moments and involve variable-dependent tuning parameters, which hamper their fidelity in applications to large-scale problems. To liberate these constraints, this paper revisits the celebrated Hodges-Lehmann (HL) estimator for estimating location parameters in both the one- and two-sample problems, from a nonasymptotic perspective. Our study develops Berry-Esseen inequality and Cramer-type moderate deviation for the HL estimator based on newly developed nonasymptotic Bahadur representation and builds data-driven confidence intervals via a weighted bootstrap approach. These results allow us to extend the HL estimator to large-scale studies and propose tuning-free and moment-free high-dimensional inference procedures for testing global null and for large-scale multiple testing with false discovery proportion control. It is convincingly shown that the resulting tuning-free and moment-free methods control false discovery proportion at a prescribed level. The simulation studies lend further support to our developed theory.
引用
下载
收藏
页码:2093 / 2115
页数:23
相关论文
共 50 条
  • [21] On robust regression with high-dimensional predictors
    El Karoui, Noureddine
    Bean, Derek
    Bickel, Peter J.
    Lim, Chinghway
    Yu, Bin
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2013, 110 (36) : 14557 - 14562
  • [22] Testing covariates in high-dimensional regression
    Lan, Wei
    Wang, Hansheng
    Tsai, Chih-Ling
    ANNALS OF THE INSTITUTE OF STATISTICAL MATHEMATICS, 2014, 66 (02) : 279 - 301
  • [23] Testing covariates in high-dimensional regression
    Wei Lan
    Hansheng Wang
    Chih-Ling Tsai
    Annals of the Institute of Statistical Mathematics, 2014, 66 : 279 - 301
  • [24] Testing for heteroscedasticity in high-dimensional regressions
    Li, Zhaoyuan
    Yao, Jianfeng
    ECONOMETRICS AND STATISTICS, 2019, 9 : 122 - 139
  • [25] Power in High-Dimensional Testing Problems
    Kock, Anders Bredahl
    Preinerstorfer, David
    ECONOMETRICA, 2019, 87 (03) : 1055 - 1069
  • [26] Testing Monotone High-Dimensional Distributions
    Rubinfeld, Ronitt
    Servedio, Rocco A.
    RANDOM STRUCTURES & ALGORITHMS, 2009, 34 (01) : 24 - 44
  • [27] ON TESTING FOR HIGH-DIMENSIONAL WHITE NOISE
    Li, Zeng
    Lam, Clifford
    Yao, Jianfeng
    Yao, Qiwei
    ANNALS OF STATISTICS, 2019, 47 (06): : 3382 - 3412
  • [28] TESTING IN HIGH-DIMENSIONAL SPIKED MODELS
    Johnstone, Iain M.
    Onatski, Alexei
    ANNALS OF STATISTICS, 2020, 48 (03): : 1231 - 1254
  • [29] A strong converse bound for multiple hypothesis testing, with applications to high-dimensional estimation
    Venkataramanan, Ramji
    Johnson, Oliver
    ELECTRONIC JOURNAL OF STATISTICS, 2018, 12 (01): : 1126 - 1149
  • [30] Streaming Algorithms for High-Dimensional Robust Statistics
    Diakonikolas, Ilias
    Kane, Daniel M.
    Pensia, Ankit
    Pittas, Thanasis
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,