Uniform Sampling for Matrix Approximation

被引:93
|
作者
Cohen, Michael B. [1 ]
Lee, Yin Tat
Musco, Cameron
Musco, Christopher
Peng, Richard
Sidford, Aaron
机构
[1] MIT, Dept EECS, 77 Massachusetts Ave, Cambridge, MA 02139 USA
来源
PROCEEDINGS OF THE 6TH INNOVATIONS IN THEORETICAL COMPUTER SCIENCE (ITCS'15) | 2015年
基金
美国国家科学基金会;
关键词
Regression; leverage scores; matrix sampling; randomized numerical linear algebra; MONTE-CARLO ALGORITHMS; FAST RANDOMIZED ALGORITHM;
D O I
10.1145/2688073.2688113
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Random sampling has become a critical tool in solving massive matrix problems. For linear regression, a small, manageable set of data rows can be randomly selected to approximate a tall, skinny data matrix, improving processing time significantly. For theoretical performance guarantees, each row must be sampled with probability proportional to its statistical leverage score. Unfortunately, leverage scores are difficult to compute. A simple alternative is to sample rows uniformly at random. While this often works, uniform sampling will eliminate critical row information for many natural instances. We take a fresh look at uniform sampling by examining what information it does preserve. Specifically, we show that uniform sampling yields a matrix that, in some sense, well approximates a large fraction of the original. While this weak form of approximation is not enough for solving linear regression directly, it is enough to compute a better approximation. This observation leads to simple iterative row sampling algorithms for matrix approximation that run in input-sparsity time and preserve row structure and sparsity at all intermediate steps. In addition to an improved understanding of uniform sampling, our main proof introduces a structural result of independent interest: we show that every matrix can be made to have low coherence by reweighting a small subset of its rows.
引用
收藏
页码:181 / 190
页数:10
相关论文
共 50 条