On Expected Error of Randomized Nystrom Kernel Regression

被引:0
|
作者
Trokicic, Aleksandar [1 ]
Todorovic, Branimir [1 ]
机构
[1] Univ Nis, Fac Sci & Math, Dept Comp Sci, Visegradska 33, Nish 18000, Serbia
关键词
kernel regression; kernel matrix; Nystrom method; randomized svd; random features;
D O I
10.2298/FIL2011871T
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Kernel methods are a class of machine learning algorithms which learn and discover patterns in a high (possibly infinite) dimensional feature space obtained by often nonlinear, possibly infinite mapping of an input space. A major problem with kernel methods is their time complexity. For a data set with n input points a time complexity of a kernel method is O(n(3)), which is intractable for a large data set. A method based on a random Nystrom features is an approximation method that is able to reduce the time complexity to O(np(2) + p(3)) where p is the number of randomly selected input data points. A time complexity of O(p(3)) comes from the fact that a spectral decomposition needs to be performed on a p x p Gram matrix, and if p is a large number even an approximate algorithm is time consuming. In this paper we will apply the randomized SVD method instead of the spectral decomposition and further reduce the time complexity. An input parameters of a randomized SVD algorithm are p < p Gram matrix and a number m < p. In this case time complexity is O(nm(2) + p(2)m + m(3)), and linear regression is performed on a m-dimensional random features. We will prove that the error of a predictor, learned via this method is almost the same in expectation as the error of a kernel predictor. Aditionally, we will empirically show that this predictor is better than the ONE that uses only Nystrom method.
引用
收藏
页码:3871 / 3884
页数:14
相关论文
共 50 条
  • [31] Error analysis of regularized least-square regression with Fredholm kernel
    Tao, Yanfang
    Yuan, Peipei
    Song, Biqin
    [J]. NEUROCOMPUTING, 2017, 249 : 237 - 244
  • [32] Superconvergent Nystrom and degenerate kernel methods for eigenvalue problems
    Alouch, C.
    Sablonniere, P.
    Sbibih, D.
    Tahrichi, M.
    [J]. APPLIED MATHEMATICS AND COMPUTATION, 2011, 217 (20) : 7851 - 7866
  • [33] Improved Bounds for the Nystrom Method With Application to Kernel Classification
    Jin, Rong
    Yang, Tianbao
    Mahdavi, Mehrdad
    Li, Yu-Feng
    Zhou, Zhi-Hua
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2013, 59 (10) : 6939 - 6949
  • [34] Kernel Least Mean Square Based on the Nystrom Method
    Wang, Shi-Yuan
    Wang, Wen-Yue
    Dang, Lu-Juan
    Jiang, Yun-Xiang
    [J]. CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2019, 38 (07) : 3133 - 3151
  • [35] Nystrom-based approximate kernel subspace learning
    Iosifidis, Alexandros
    Gabbouj, Moncef
    [J]. PATTERN RECOGNITION, 2016, 57 : 190 - 197
  • [36] Using the Nystrom method to speed up kernel machines
    Williams, CKI
    Seeger, M
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 13, 2001, 13 : 682 - 688
  • [37] A Linear Incremental Nystrom Method for Online Kernel Learning
    Xu, Shan
    Zhang, Xiao
    Liao, Shizhong
    [J]. 2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 2256 - 2261
  • [38] Kernel K-Means Sampling for Nystrom Approximation
    He, Li
    Zhang, Hong
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (05) : 2108 - 2120
  • [39] Fast DPP Sampling for Nystrom with Application to Kernel Methods
    Li, Chengtao
    Jegelka, Stefanie
    Sra, Suvrit
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [40] Fast and Accurate Refined Nystrom Based Kernel SVM
    Li, Zhe
    Yang, Tianbao
    Zhang, Lijun
    Jin, Rong
    [J]. THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1830 - 1836