Non-asymptotic analysis and inference for an outlyingness induced winsorized mean

被引:0
|
作者
Zuo, Yijun [1 ]
机构
[1] Michigan State Univ, Dept Stat & Probabil, E Lansing, MI 48824 USA
关键词
Non-asymptotic analysis; Centrality estimation; Sub-Gaussian performance; Computability; Finite sample breakdown point; MULTIVARIATE LOCATION; DEPTH; COVARIANCE;
D O I
10.1007/s00362-022-01353-5
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Robust estimation of a mean vector, a topic regarded as obsolete in the traditional robust statistics community, has recently surged in machine learning literature in the last decade. The latest focus is on the sub-Gaussian performance and computability of the estimators in a non-asymptotic setting. Numerous traditional robust estimators are computationally intractable, which partly contributes to the renewal of the interest in the robust mean estimation. Robust centrality estimators, however, include the trimmed mean and the sample median. The latter has the best robustness but suffers a low efficiency drawback. Trimmed mean and median of means, achieving sub-Gaussian performance have been proposed and studied in the literature. This article investigates the robustness of leading sub-Gaussian estimators of mean and reveals that none of them can resist greater than 25% contamination in data and consequently introduces an outlyingness induced winsorized mean which has the best possible robustness (can resist up to 50% contamination without breakdown) meanwhile achieving high efficiency. Furthermore, it has a sub-Gaussian performance for uncontaminated samples and a bounded estimation error for contaminated samples at a given confidence level in a finite sample setting. It can be computed in linear time.
引用
收藏
页码:1465 / 1481
页数:17
相关论文
共 50 条
  • [1] Non-asymptotic analysis and inference for an outlyingness induced winsorized mean
    Yijun Zuo
    Statistical Papers, 2023, 64 : 1465 - 1481
  • [2] Non-asymptotic analysis of tangent space perturbation
    Kaslovsky, Daniel N.
    Meyer, Francois G.
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2014, 3 (02) : 134 - 187
  • [3] Non-asymptotic calibration and resolution
    Vovk, V
    ALGORITHMIC LEARNING THEORY, 2005, 3734 : 429 - 443
  • [4] Non-Asymptotic Analysis for Relational Learning with One Network
    He, Peng
    Zhang, Changshui
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 33, 2014, 33 : 320 - 327
  • [5] Feedback in the Non-Asymptotic Regime
    Polyanskiy, Yury
    Poor, H. Vincent
    Verdu, Sergio
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2011, 57 (08) : 4903 - 4925
  • [6] Non-asymptotic Analysis of Biased Stochastic Approximation Scheme
    Karimi, Belhal
    Miasojedow, Blazej
    Moulines, Eric
    Wai, Hoi-To
    CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [7] Testing for Anomalies: Active Strategies and Non-asymptotic Analysis
    Kartik, Dhruva
    Nayyar, Ashutosh
    Mitra, Urbashi
    2020 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2020, : 1277 - 1282
  • [8] Language Approximation: Asymptotic and Non-asymptotic Results
    Ravikumar, Bala
    DEVELOPMENTS IN LANGUAGE THEORY, DLT 2017, 2017, 10396
  • [9] Non-Asymptotic Mean and Variance Also Approximately Satisfy Taylor's Law
    Cohen, Michael P.
    AMERICAN STATISTICIAN, 2017, 71 (02): : 187 - 187
  • [10] Non-Asymptotic Analysis of Monte Carlo Tree Search
    Shah D.
    Xie Q.
    Xu Z.
    Performance Evaluation Review, 2020, 48 (01): : 31 - 32