Diversity Sampling is an Implicit Regularization for Kernel Methods

被引:8
|
作者
Fanuel, Michael [1 ]
Schreurs, Joachim [1 ]
Suykens, Johan A. K. [1 ]
机构
[1] Katholieke Univ Leuven, Dept Elect Engn ESAT, STADIUS Ctr Dynam Syst, Signal Proc & Data Analyt, B-3001 Leuven, Belgium
来源
基金
欧洲研究理事会;
关键词
kernel; Nystrom; determinantal point process; regularization; NYSTROM;
D O I
10.1137/20M1320031
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Kernel methods have achieved very good performance on large scale regression and classification problems by using the Nystrom method and preconditioning techniques. The Nystrom approximation--based on a subset of landmarks-gives a low rank approximation of the kernel matrix, and is known to provide a form of implicit regularization. We further elaborate on the impact of sampling diverse landmarks for constructing the Nystrom approximation in supervised as well as unsupervised kernel methods. By using Determinantal Point Processes (DPP) for sampling, we obtain additional theoretical results concerning the interplay between diversity and regularization. Empirically, we demonstrate the advantages of training kernel methods based on subsets made of diverse points. In particular, if the dataset has a dense bulk and a sparser tail, we show that Nystrom kernel regression with diverse landmarks increases the accuracy of the regression in sparser regions of the dataset, with respect to a uniform landmark sampling. A greedy heuristic is also proposed to select diverse samples of significant size within large datasets when exact DPP sampling is not practically feasible.
引用
收藏
页码:280 / 297
页数:18
相关论文
共 50 条
  • [1] Sampling techniques for kernel methods
    Achlioptas, D
    McSherry, F
    Schölkopf, B
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 14, VOLS 1 AND 2, 2002, 14 : 335 - 342
  • [2] Implicit Regularization of Accelerated Methods in Hilbert Spaces
    Pagliana, Nicolo
    Rosasco, Lorenzo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Near-Orthogonality Regularization in Kernel Methods
    Xie, Pengtao
    Poczos, Barnabas
    Xing, Eric P.
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017), 2017,
  • [4] New regularization methods for convolutional kernel tensors
    Guo, Pei-Chang
    AIMS MATHEMATICS, 2023, 8 (11): : 26188 - 26198
  • [5] An Alternative to Tikhonov Regularization for Linear Sampling Methods
    Kim, K.
    Leem, K. H.
    Pelekanos, G.
    ACTA APPLICANDAE MATHEMATICAE, 2010, 112 (02) : 171 - 180
  • [6] An Alternative to Tikhonov Regularization for Linear Sampling Methods
    K. Kim
    K. H. Leem
    G. Pelekanos
    Acta Applicandae Mathematicae, 2010, 112 : 171 - 180
  • [7] A practical use of regularization for supervised learning with kernel methods
    Prato, M.
    Zanni, L.
    PATTERN RECOGNITION LETTERS, 2013, 34 (06) : 610 - 618
  • [8] An implicit gradient model by a reproducing kernel strain regularization in strain localization problems
    Chen, JS
    Zhang, XW
    Belytschko, T
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2004, 193 (27-29) : 2827 - 2844
  • [9] A Unified Framework for Discrete Multi-kernel k-means with Kernel Diversity Regularization
    Lu, Yihang
    Zheng, Xuan
    Wang, Rong
    Nie, Feiping
    Li, Xuelong
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4934 - 4940
  • [10] Kernel methods in line and point transect sampling
    Mack, YP
    Quang, PX
    BIOMETRICS, 1998, 54 (02) : 606 - 619