Attribute-efficient learning of monomials over highly-correlated variables

被引:0
|
作者
Andoni, Alexandr [1 ,2 ]
Dudeja, Rishabh [3 ]
Hsu, Daniel [1 ,2 ]
Vodrahalli, Kiran [1 ]
机构
[1] Columbia Univ, Dept Comp Sci, New York, NY 10027 USA
[2] Columbia Univ, Data Sci Inst, New York, NY 10027 USA
[3] Columbia Univ, Dept Stat, New York, NY 10027 USA
来源
关键词
attribute-efficient; computationally efficient; statistics; monomial; learning; restricted eigenvalue condition; lasso; log-transform; dependent features;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of learning a real-valued function of correlated variables. Solving this problem is of interest since many classical learning results apply only in the case of learning functions of random variables that are independent. We show how to recover a high-dimensional, sparse monomial model from Gaussian examples with sample complexity that is poly-logarithmic in the total number of variables and polynomial in the number of relevant variables. Our algorithm is based on a transformation of the variables-taking their logarithm-followed by a sparse linear regression procedure, which is statistically and computationally efficient. While this transformation is commonly used in applied non-linear regression, its statistical guarantees have never been rigorously analyzed. We prove that the sparse regression procedure succeeds even in cases where the original features are highly correlated and fail to satisfy the standard assumptions required for sparse linear regression.
引用
收藏
页数:35
相关论文
共 15 条