A comparative study on large scale kernelized support vector machines

被引:0
|
作者
Daniel Horn
Aydın Demircioğlu
Bernd Bischl
Tobias Glasmachers
Claus Weihs
机构
[1] Technische Universität Dortmund,Fakultät Statistik
[2] Ruhr-Universität Bochum,Department of Statistics
[3] LMU München,undefined
关键词
Support vector machine; Multi-objective optimization; Supervised learning; Machine learning; Large scale; Nonlinear SVM; Parameter tuning; 62-07 Data analysis;
D O I
暂无
中图分类号
学科分类号
摘要
Kernelized support vector machines (SVMs) belong to the most widely used classification methods. However, in contrast to linear SVMs, the computation time required to train such a machine becomes a bottleneck when facing large data sets. In order to mitigate this shortcoming of kernel SVMs, many approximate training algorithms were developed. While most of these methods claim to be much faster than the state-of-the-art solver LIBSVM, a thorough comparative study is missing. We aim to fill this gap. We choose several well-known approximate SVM solvers and compare their performance on a number of large benchmark data sets. Our focus is to analyze the trade-off between prediction error and runtime for different learning and accuracy parameter settings. This includes simple subsampling of the data, the poor-man’s approach to handling large scale problems. We employ model-based multi-objective optimization, which allows us to tune the parameters of learning machine and solver over the full range of accuracy/runtime trade-offs. We analyze (differences between) solvers by studying and comparing the Pareto fronts formed by the two objectives classification error and training time. Unsurprisingly, given more runtime most solvers are able to find more accurate solutions, i.e., achieve a higher prediction accuracy. It turns out that LIBSVM with subsampling of the data is a strong baseline. Some solvers systematically outperform others, which allows us to give concrete recommendations of when to use which solver.
引用
收藏
页码:867 / 883
页数:16
相关论文
共 50 条
  • [21] Large-Scale Training of Pairwise Support Vector Machines for Speaker Recognition
    Cumani, Sandro
    Laface, Pietro
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2014, 22 (11) : 1590 - 1600
  • [22] Hash-Based Support Vector Machines Approximation for Large Scale Prediction
    Litayem, Saloua
    Joly, Alexis
    Boujemaa, Nozha
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2012, 2012,
  • [23] Large-scale training of pairwise support vector machines for speaker recognition
    Cumani, Sandro, 1600, Institute of Electrical and Electronics Engineers Inc., United States (22):
  • [24] Large-scale image retrieval using transductive support vector machines
    Cevikalp, Hakan
    Elmas, Merve
    Ozkan, Savas
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2018, 173 : 2 - 12
  • [25] Parallel software for training large scale support vector machines on multiprocessor systems
    Zanni, Luca
    Serafini, Thomas
    Zanghirati, Gaetano
    JOURNAL OF MACHINE LEARNING RESEARCH, 2006, 7 : 1467 - 1492
  • [26] A divide-and-conquer method for large scale ν-nonparallel support vector machines
    Xuchan Ju
    Yingjie Tian
    Neural Computing and Applications, 2018, 29 : 497 - 509
  • [27] A divide-and-conquer method for large scale ν-nonparallel support vector machines
    Ju, Xuchan
    Tian, Yingjie
    NEURAL COMPUTING & APPLICATIONS, 2018, 29 (09): : 497 - 509
  • [28] A Hash Based Method for Large Scale Nonparallel Support Vector Machines Prediction
    Ju, Xuchan
    Wang, Tianhe
    INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE (ICCS 2017), 2017, 108 : 1281 - 1291
  • [29] A divide-and-combine method for large scale nonparallel support vector machines
    Tian, Yingjie
    Ju, Xuchan
    Shi, Yong
    NEURAL NETWORKS, 2016, 75 : 12 - 21
  • [30] A comparative study of multi-class support vector machines in the unifying framework of large margin classifiers
    Guermeur, Y
    Elisseeff, A
    Zelus, D
    APPLIED STOCHASTIC MODELS IN BUSINESS AND INDUSTRY, 2005, 21 (02) : 199 - 214