Linear Regression With Distributed Learning: A Generalization Error Perspective

被引:5
|
作者
Hellkvist, Martin [1 ]
Ozcelikkale, Ayca [1 ]
Ahlen, Anders [1 ]
机构
[1] Uppsala Univ, Dept Elect Engn, S-75237 Uppsala, Sweden
基金
瑞典研究理事会;
关键词
Distance learning; Computer aided instruction; Training; Data models; Training data; Distributed databases; Numerical models; Distributed estimation; distributed optimization; supervised learning; generalization error; networked systems; OPTIMIZATION; ALGORITHMS;
D O I
10.1109/TSP.2021.3106441
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Distributed learning provides an attractive framework for scaling the learning task by sharing the computational load over multiple nodes in a network. Here, we investigate the performance of distributed learning for large-scale linear regression where the model parameters, i.e., the unknowns, are distributed over the network. We adopt a statistical learning approach. In contrast to works that focus on the performance on the training data, we focus on the generalization error, i.e., the performance on unseen data. We provide high-probability bounds on the generalization error for both isotropic and correlated Gaussian data as well as sub-gaussian data. These results reveal the dependence of the generalization performance on the partitioning of the model over the network. In particular, our results show that the generalization error of the distributed solution can be substantially higher than that of the centralized solution even when the error on the training data is at the same level for both the centralized and distributed approaches. Our numerical results illustrate the performance with both real-world image data as well as synthetic data.
引用
收藏
页码:5479 / 5495
页数:17
相关论文
共 50 条
  • [1] Generalization Error for Linear Regression under Distributed Learning
    Hellkvist, Martin
    Ozcelikkale, Ayca
    Ahlen, Anders
    PROCEEDINGS OF THE 21ST IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC2020), 2020,
  • [2] Active learning in approximately linear regression based on conditional expectation of generalization error
    Department of Computer Science, Tokyo Institute of Technology, 2-12-1, O-okayama, Meguro-ku, Tokyo, 152-8552, Japan
    J. Mach. Learn. Res., 2006, (141-166):
  • [3] Active learning in approximately linear regression based on conditional expectation of generalization error
    Sugiyama, M
    JOURNAL OF MACHINE LEARNING RESEARCH, 2006, 7 : 141 - 166
  • [4] Generalization error estimation for non-linear learning methods
    Sugiyama, Masashi
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2007, E90A (07) : 1496 - 1499
  • [5] ANOTHER PERSPECTIVE ON LINEAR REGRESSION, STRUCTURAL RELATIONS, AND MEASUREMENT ERROR
    WERTS, CE
    LINN, RL
    JORESKOG, KG
    EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1973, 33 (02) : 327 - 332
  • [6] Rate-Distortion Theoretic Bounds on Generalization Error for Distributed Learning
    Sefidgaran, Milad
    Chor, Romain
    Zaidi, Abdellatif
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [7] Fast generalization error bound of deep learning from a kernel perspective
    Suzuki, Taiji
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [8] Regression with Deep Neural Networks: Generalization Error Guarantees, Learning Algorithms, and Regularizers
    Amjad, Jaweria
    Lyu, Zhaoyan
    Rodrigues, Miguel R. D.
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1481 - 1485
  • [9] Generalization Error in Deep Learning
    Jakubovitz, Daniel
    Giryes, Raja
    Rodrigues, Miguel R. D.
    COMPRESSED SENSING AND ITS APPLICATIONS, 2019, : 153 - 193
  • [10] GENERALIZATION ERROR BOUNDS OF DYNAMIC TREATMENT REGIMES IN PENALIZED REGRESSION-BASED LEARNING
    Oh, Eun Jeong
    Qian, Min
    Cheung, Ying Kuen
    ANNALS OF STATISTICS, 2022, 50 (04): : 2047 - 2071