Truncated Linear Regression in High Dimensions

被引:0
|
作者
Daskalakis, Constantinos [1 ]
Rohatgi, Dhruv [1 ]
Zampetakis, Manolis [1 ]
机构
[1] MIT, Cambridge, MA 02139 USA
关键词
RECOVERY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As in standard linear regression, in truncated linear regression, we are given access to observations (A(i), y(i))(i) whose dependent variable equals y(i) = A(t)(T). x* + eta(i), where x* is some fixed unknown vector of interest and eta(i) is independent noise; except we are only given an observation if its dependent variable y(i) lies in some "truncation set" S subset of R. The goal is to recover x* under some favorable conditions on the A(i)'s and the noise distribution. We prove that there exists a computationally and statistically efficient method for recovering k-sparse n-dimensional vectors x* from m truncated samples, which attains an optimal (2) pound reconstruction error of O(J(k log n)/m). As a corollary, our guarantees imply a computationally efficient and information-theoretically optimal algorithm for compressed sensing with truncation, which may arise from measurement saturation effects. Our result follows from a statistical and computational analysis of the Stochastic Gradient Descent (SGD) algorithm for solving a natural adaptation of the LASSO optimization problem that accommodates truncation. This generalizes the works of both: (1) Daskalakis et al. [9], where no regularization is needed due to the low-dimensionality of the data, and (2) Wainright [27], where the objective function is simple due to the absence of truncation. In order to deal with both truncation and high-dimensionality at the same time, we develop new techniques that not only generalize the existing ones but we believe are of independent interest.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Strategyproof Linear Regression in High Dimensions
    Chen, Yiling
    Podimata, Chara
    Procaccia, Ariel D.
    Shah, Nisarg
    [J]. ACM EC'18: PROCEEDINGS OF THE 2018 ACM CONFERENCE ON ECONOMICS AND COMPUTATION, 2018, : 9 - 26
  • [2] Partially functional linear regression in high dimensions
    Kong, Dehan
    Xue, Kaijie
    Yao, Fang
    Zhang, Hao H.
    [J]. BIOMETRIKA, 2016, 103 (01) : 147 - 159
  • [3] Strategyproof Linear Regression in High Dimensions: An Overview
    Chen, Yiling
    Podimata, Chara
    Procaccia, Ariel D.
    Shah, Nisarg
    [J]. ACM SIGECOM EXCHANGES, 2019, 17 (01) : 54 - 60
  • [4] Functional linear regression with truncated signatures
    Fermanian, Adeline
    [J]. JOURNAL OF MULTIVARIATE ANALYSIS, 2022, 192
  • [5] NONNORMAL LINEAR-REGRESSION - AN EXAMPLE OF SIGNIFICANCE LEVELS IN HIGH DIMENSIONS
    FRASER, DAS
    LEE, HS
    REID, N
    [J]. BIOMETRIKA, 1990, 77 (02) : 333 - 341
  • [6] Smoothed quantile regression for partially functional linear models in high dimensions
    Wang, Zhihao
    Bai, Yongxin
    Haerdle, Wolfgang K.
    Tian, Maozai
    [J]. BIOMETRICAL JOURNAL, 2023, 65 (07)
  • [7] Local linear quantile regression with truncated and dependent data
    Wang, Jiang-Feng
    Ma, Wei-Min
    Fan, Guo-Liang
    Wen, Li-Min
    [J]. STATISTICS & PROBABILITY LETTERS, 2015, 96 : 232 - 240
  • [8] Efficient Truncated Linear Regression with Unknown Noise Variance
    Daskalakis, Constantinos
    Stefanou, Patroklos
    Yao, Rui
    Zampetakis, Manolis
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [9] Truncated estimation in functional generalized linear regression models
    Liu, Xi
    Divani, Afshin A.
    Petersen, Alexander
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2022, 169
  • [10] Weighted estimation in linear regression for truncated survival data
    Gross, ST
    [J]. SCANDINAVIAN JOURNAL OF STATISTICS, 1996, 23 (02) : 179 - 193