Communication-efficient distributed estimation for high-dimensional large-scale linear regression

被引:0
|
作者
Zhan Liu
Xiaoluo Zhao
Yingli Pan
机构
[1] Hubei University,Hubei Key Laboratory of Applied Mathematics, Faculty of Mathematics and Statistics
来源
Metrika | 2023年 / 86卷
关键词
Distributed optimization; SCAD; Adaptive LASSO; GEL function; Modified proximal ADMM algorithm;
D O I
暂无
中图分类号
学科分类号
摘要
In the Master-Worker distributed structure, this paper provides a regularized gradient-enhanced loss (GEL) function based on the high-dimensional large-scale linear regression with SCAD and adaptive LASSO penalty. The importance and originality of this paper have two aspects: (1) Computationally, to take full advantage of the computing power of each machine and speed up the convergence, our proposed distributed upgraded estimation method can make all Workers optimize their corresponding GEL functions in parallel, and the results are then aggregated by the Master; (2) In terms of communication, the proposed modified proximal alternating direction method of the multipliers (ADMM) algorithm is comparable to the Centralize method based on the full sample during a few rounds of communication. Under some mild assumptions, we establish the Oracle properties of the SCAD and adaptive LASSO penalized linear regression. The finite sample properties of the newly suggested method are assessed through simulation studies. An application to the HIV drug susceptibility study demonstrates the utility of the proposed method in practice.
引用
收藏
页码:455 / 485
页数:30
相关论文
共 50 条
  • [1] Communication-efficient distributed estimation for high-dimensional large-scale linear regression
    Liu, Zhan
    Zhao, Xiaoluo
    Pan, Yingli
    [J]. METRIKA, 2023, 86 (04) : 455 - 485
  • [2] Communication-efficient estimation of high-dimensional quantile regression
    Wang, Lei
    Lian, Heng
    [J]. ANALYSIS AND APPLICATIONS, 2020, 18 (06) : 1057 - 1075
  • [3] Communication-efficient distributed estimation of partially linear additive models for large-scale data
    Gao, Junzhuo
    Wang, Lei
    [J]. INFORMATION SCIENCES, 2023, 631 : 185 - 201
  • [4] Communication-efficient distributed large-scale sparse multinomial logistic regression
    Lei, Dajiang
    Huang, Jie
    Chen, Hao
    Li, Jie
    Wu, Yu
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2023, 35 (18):
  • [5] Communication-Efficient Distributed Estimation of Causal Effects With High-Dimensional Data
    Wang, Xiaohan
    Tong, Jiayi
    Peng, Sida
    Chen, Yong
    Ning, Yang
    [J]. STAT, 2024, 13 (03):
  • [6] Efficient distributed optimization for large-scale high-dimensional sparse penalized Huber regression
    Pan, Yingli
    Xu, Kaidong
    Wei, Sha
    Wang, Xiaojuan
    Liu, Zhan
    [J]. COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2024, 53 (07) : 3106 - 3125
  • [7] Communication-Efficient Distributed Multiple Testing for Large-Scale Inference
    Pournaderi, Mehrdad
    Xiang, Yu
    [J]. arXiv, 2022,
  • [8] Communication-efficient estimation and inference for high-dimensional quantile regression based on smoothed decorrelated score
    Di, Fengrui
    Wang, Lei
    Lian, Heng
    [J]. STATISTICS IN MEDICINE, 2022, 41 (25) : 5084 - 5101
  • [9] Communication-Efficient Distributed Learning for High-Dimensional Support Vector Machines
    Zhou, Xingcai
    Shen, Hao
    [J]. MATHEMATICS, 2022, 10 (07)
  • [10] Communication-efficient low-dimensional parameter estimation and inference for high-dimensional LP-quantile regression
    Gao, Junzhuo
    Wang, Lei
    [J]. SCANDINAVIAN JOURNAL OF STATISTICS, 2024, 51 (01) : 302 - 333