Distributed Continual Learning With CoCoA in High-Dimensional Linear Regression

被引:0
|
作者
Hellkvist, Martin [1 ]
Ozcelikkale, Ayca [1 ]
Ahlen, Anders [1 ]
机构
[1] Uppsala Univ, Dept Elect Engn, S-75121 Uppsala, Sweden
基金
瑞典研究理事会;
关键词
Task analysis; Training; Distributed databases; Distance learning; Computer aided instruction; Data models; Training data; Multi-task networks; networked systems; distributed estimation; adaptation; overparametrization; NEURAL-NETWORKS; ALGORITHMS;
D O I
10.1109/TSP.2024.3361714
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We consider estimation under scenarios where the signals of interest exhibit change of characteristics over time. In particular, we consider the continual learning problem where different tasks, e.g., data with different distributions, arrive sequentially and the aim is to perform well on the newly arrived task without performance degradation on the previously seen tasks. In contrast to the continual learning literature focusing on the centralized setting, we investigate the problem from a distributed estimation perspective. We consider the well-established distributed learning algorithm CoCoA, which distributes the model parameters and the corresponding features over the network. We provide exact analytical characterization for the generalization error of CoCoA under continual learning for linear regression in a range of scenarios, where overparameterization is of particular interest. These analytical results characterize how the generalization error depends on the network structure, the task similarity and the number of tasks, and show how these dependencies are intertwined. In particular, our results show that the generalization error can be significantly reduced by adjusting the network size, where the most favorable network size depends on task similarity and the number of tasks. We present numerical results verifying the theoretical analysis and illustrate the continual learning performance of CoCoA with a digit classification task.
引用
收藏
页码:1015 / 1031
页数:17
相关论文
共 50 条
  • [21] Spline-Lasso in High-Dimensional Linear Regression
    Guo, Jianhua
    Hu, Jianchang
    Jing, Bing-Yi
    Zhang, Zhen
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2016, 111 (513) : 288 - 297
  • [22] CANONICAL THRESHOLDING FOR NONSPARSE HIGH-DIMENSIONAL LINEAR REGRESSION
    Silin, Igor
    Fan, Jianqing
    ANNALS OF STATISTICS, 2022, 50 (01): : 460 - 486
  • [23] Robust Estimation of High-Dimensional Linear Regression with Changepoints
    Cui X.
    Geng H.
    Wang Z.
    Zou C.
    IEEE Transactions on Information Theory, 2024, 70 (10) : 1 - 1
  • [24] Robust linear regression for high-dimensional data: An overview
    Filzmoser, Peter
    Nordhausen, Klaus
    WILEY INTERDISCIPLINARY REVIEWS-COMPUTATIONAL STATISTICS, 2021, 13 (04)
  • [25] A global homogeneity test for high-dimensional linear regression
    Charbonnier, Camille
    Verzelen, Nicolas
    Villers, Fanny
    ELECTRONIC JOURNAL OF STATISTICS, 2015, 9 (01): : 318 - 382
  • [26] Empirical likelihood for high-dimensional linear regression models
    Hong Guo
    Changliang Zou
    Zhaojun Wang
    Bin Chen
    Metrika, 2014, 77 : 921 - 945
  • [27] High-dimensional analysis of variance in multivariate linear regression
    Lou, Zhipeng
    Zhang, Xianyang
    Wu, Wei Biao
    BIOMETRIKA, 2023, 110 (03) : 777 - 797
  • [28] THE TAP FREE ENERGY FOR HIGH-DIMENSIONAL LINEAR REGRESSION
    Qiu, Jiaze
    Sen, Subhabrata
    ANNALS OF APPLIED PROBABILITY, 2023, 33 (04): : 2643 - 2680
  • [29] Leveraging independence in high-dimensional mixed linear regression
    Wang, Ning
    Deng, Kai
    Mai, Qing
    Zhang, Xin
    BIOMETRICS, 2024, 80 (03)
  • [30] Empirical likelihood for high-dimensional linear regression models
    Guo, Hong
    Zou, Changliang
    Wang, Zhaojun
    Chen, Bin
    METRIKA, 2014, 77 (07) : 921 - 945