Distributed adaptive Huber regression

被引:9
|
作者
Luo, Jiyu [1 ]
Sun, Qiang [2 ]
Zhou, Wen-Xin [3 ]
机构
[1] Univ Calif San Diego, Herbert Wertheim Sch Publ Hlth & Human Longev Sci, Div Biostat, San Diego, CA 92093 USA
[2] Univ Toronto, Dept Stat Sci, Toronto, ON M5S 3G3, Canada
[3] Univ Calif San Diego, Dept Math, La Jolla, CA 92093 USA
基金
加拿大自然科学与工程研究理事会; 美国国家科学基金会;
关键词
Adaptive Huber regression; Communication efficiency; Distributed inference; Heavy-tailed distribution; Nonasymptotic analysis; ROBUST REGRESSION; QUANTILE REGRESSION; M-ESTIMATORS; ASYMPTOTIC-BEHAVIOR; LINEAR-REGRESSION; PARAMETERS;
D O I
10.1016/j.csda.2021.107419
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Distributed data naturally arise in scenarios involving multiple sources of observations, each stored at a different location. Directly pooling all the data together is often prohibited due to limited bandwidth and storage, or due to privacy protocols. A new robust distributed algorithm is introduced for fitting linear regressions when data are subject to heavy-tailed and/or asymmetric errors with finite second moments. The algorithm only communicates gradient information at each iteration, and therefore is communication-efficient. To achieve the bias-robustness tradeoff, the key is a novel double-robustification approach that applies on both the local and global objective functions. Statistically, the resulting estimator achieves the centralized nonasymptotic error bound as if all the data were pooled together and came from a distribution with sub-Gaussian tails. Under a finite (2 + delta)-th moment condition, a Berry-Esseen bound for the distributed estimator is established, based on which robust confidence intervals are constructed. In high dimensions, the proposed doubly-robustified loss function is complemented with l(1) -penalization for fitting sparse linear models with distributed data. Numerical studies further confirm that compared with extant distributed methods, the proposed methods achieve near-optimal accuracy with low variability and better coverage with tighter confidence width. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:23
相关论文
共 50 条
  • [21] LOW-RANK TENSOR HUBER REGRESSION
    Wei, Yangxin
    Luot, Ziyan
    Chen, Yang
    PACIFIC JOURNAL OF OPTIMIZATION, 2022, 18 (02): : 439 - 458
  • [22] High-Dimensional Constrained Huber Regression
    Wei, Quan
    Zhao, Ziping
    2024 IEEE 13RD SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP, SAM 2024, 2024,
  • [23] A Semismooth Newton Method for Adaptive Distributed Sparse Linear Regression
    Shutin, Dmitriy
    Vexler, Boris
    2015 IEEE 6TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP), 2015, : 433 - 436
  • [24] Safe feature screening rules for the regularized Huber regression
    Chen, Huangyue
    Kong, Lingchen
    Shang, Pan
    Pan, Shanshan
    APPLIED MATHEMATICS AND COMPUTATION, 2020, 386
  • [25] A NEW PRINCIPLE FOR TUNING-FREE HUBER REGRESSION
    Wang, Lili
    Zheng, Chao
    Zhou, Wen
    Zhou, Wen-Xin
    STATISTICA SINICA, 2021, 31 (04) : 2153 - 2177
  • [26] ON A CONJECTURE OF HUBER CONCERNING THE CONVERGENCE OF PROJECTION PURSUIT REGRESSION
    JONES, LK
    ANNALS OF STATISTICS, 1987, 15 (02): : 880 - 882
  • [27] Uncertain regression model based on Huber loss function
    Xie, Wenxuan
    Wu, Jiali
    Sheng, Yuhong
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (01) : 1169 - 1178
  • [28] Sparse Reduced Rank Huber Regression in High Dimensions
    Tan, Kean Ming
    Sun, Qiang
    Witten, Daniela
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2023, 118 (544) : 2383 - 2393
  • [29] Heavy-tailed Linear Bandit with Huber Regression
    Kang, Minhyun
    Kim, Gi-Soo
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1027 - 1036
  • [30] Huber Regression Analysis with a Semi-Supervised Method
    Wang, Yue
    Wang, Baobin
    Peng, Chaoquan
    Li, Xuefeng
    Yin, Hong
    MATHEMATICS, 2022, 10 (20)