On using prototype reduction schemes and classifier fusion strategies to optimize kernel-based nonlinear subspace methods

被引:0
|
作者
Kim, SW [1 ]
Oommen, BJ
机构
[1] Myongji Univ, Dept Comp Sci & Engn, Yongin 449728, South Korea
[2] Carleton Univ, Sch Comp Sci, Ottawa, ON K1S 5B6, Canada
关键词
Kernel Principal Component Analysis (kPCA); kernel-based nonlinear subspace (KNS) method; prototype reduction schemes (PRS); classifier fusion strategies (CFS);
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In Kernel-based Nonlinear Subspace (KNS) methods, the length of the projections onto the principal component directions in the feature space, is computed using a kernel matrix, K, whose dimension is equivalent to the number of sample data points. Clearly this is problematic, especially, for large data sets. In this paper, we solve this problem by subdividing the data into smaller subsets, and utilizing a Prototype Reduction Scheme (PRS) as a preprocessing module, to yield more refined representative prototypes. Thereafter, a Classifier Fusion Strategy (CFS) is invoked as a postprocessing module, to combine the individual KNS classification results to derive a consensus decision. Essentially, the PRS is used to yield computational advantage, and the CFS, in turn, is used to compensate for the decreased efficiency caused by the data set division. Our experimental results demonstrate that the proposed mechanism significantly reduces the prototype extraction time as well as the computation time without sacrificing the classification accuracy. The results especially demonstrate a significant computational advantage for large data sets within a parallel processing philosophy.
引用
收藏
页数:6
相关论文
共 50 条