GRASSMANNIAN DIFFUSION MAPS-BASED DIMENSION REDUCTION AND CLASSIFICATION FOR HIGH-DIMENSIONAL DATA

被引:8
|
作者
Dos Santos, Ketson R. [1 ]
Giovanis, Dimitrios G. [1 ]
Shields, Michael D. [1 ]
机构
[1] Johns Hopkins Univ, Dept Civil & Syst Engn, Baltimore, MD 21218 USA
来源
SIAM JOURNAL ON SCIENTIFIC COMPUTING | 2022年 / 44卷 / 02期
关键词
Grassmann manifold; diffusion maps; dimension reduction; data classification; face recognition; SUBSPACES; EIGENMAPS;
D O I
10.1137/20M137001X
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This work introduces the Grassmannian diffusion maps (GDMaps), a novel nonlinear dimensionality reduction technique that defines the affinity between points through their representation as low-dimensional subspaces corresponding to points on the Grassmann manifold. The method is designed for applications, such as image recognition and data-based classification of constrained high-dimensional data where each data point itself is a high-dimensional object (i.e., a large matrix) that can be compactly represented in a lower-dimensional subspace. The GDMaps is composed of two stages. The first is a pointwise linear dimensionality reduction wherein each high-dimensional object is mapped onto the Grassmann manifold representing the low-dimensional subspace on which it resides. The second stage is a multipoint nonlinear kernel-based dimension reduction using diffusion maps to identify the subspace structure of the points on the Grassmann manifold. To this end, an appropriate Grassmannian kernel is used to construct the transition matrix of a random walk on a graph connecting points on the Grassmann manifold. Spectral analysis of the transition matrix yields low-dimensional Grassmannian diffusion coordinates embedding the data into a low-dimensional reproducing kernel Hilbert space. Further, a novel data classification/recognition technique is developed based on the construction of an overcomplete dictionary of reduced dimension whose atoms are given by the Grassmannian diffusion coordinates. Three examples are considered. First, a "toy" example shows that the GDMaps can identify an appropriate parametrization of structured points on the unit sphere. The second example demonstrates the ability of the GDMaps to revealing the intrinsic subspace structure of high-dimensional random field data. In the last example, a face recognition problem is solved considering face images subject to varying illumination conditions, changes in face expressions, and occurrence of occlusions. The technique presented high recognition rates (i.e., 95% in the best case) using a fraction of the data required by conventional methods.
引用
收藏
页码:B250 / B274
页数:25
相关论文
共 50 条
  • [1] High-dimensional Data Dimension Reduction Based on KECA
    Hu, Yongde
    Pan, Jingchang
    Tan, Xin
    SENSORS, MEASUREMENT AND INTELLIGENT MATERIALS, PTS 1-4, 2013, 303-306 : 1101 - 1104
  • [2] A Hybrid Dimension Reduction Based Linear Discriminant Analysis for Classification of High-Dimensional Data
    Zorarpaci, Ezgi
    2021 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC 2021), 2021, : 1028 - 1036
  • [3] A novel dimension reduction and dictionary learning framework for high-dimensional data classification
    Li, Yanxia
    Chai, Yi
    Zhou, Han
    Yin, Hongpeng
    PATTERN RECOGNITION, 2021, 112
  • [4] High-dimensional Data Classification Based on Principal Component Analysis Dimension Reduction and Improved BP Algorithm
    Yan, Tai-shan
    Wen, Yi-ting
    Li, Wen-bin
    2018 INTERNATIONAL CONFERENCE ON COMMUNICATION, NETWORK AND ARTIFICIAL INTELLIGENCE (CNAI 2018), 2018, : 441 - 445
  • [5] Visualisation and dimension reduction of high-dimensional data for damage detection
    Worden, K
    Manson, G
    IMAC - PROCEEDINGS OF THE 17TH INTERNATIONAL MODAL ANALYSIS CONFERENCE, VOLS I AND II, 1999, 3727 : 1576 - 1585
  • [6] DECIDING THE DIMENSION OF EFFECTIVE DIMENSION REDUCTION SPACE FOR FUNCTIONAL AND HIGH-DIMENSIONAL DATA
    Li, Yehua
    Hsing, Tailen
    ANNALS OF STATISTICS, 2010, 38 (05): : 3028 - 3062
  • [7] Efficient dimension reduction for high-dimensional matrix-valued data
    Wang, Dong
    Shen, Haipeng
    Truong, Young
    NEUROCOMPUTING, 2016, 190 : 25 - 34
  • [8] Fusion of effective dimension reduction and discriminative dictionary learning for high-dimensional classification
    Wang, Shuang-xi
    Ge, Hong-wei
    Gou, Jian-ping
    Ou, Wei-hua
    Yin, He-feng
    Su, Shu-zhi
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 99
  • [9] Dependence maps, a dimensionality reduction with dependence distance for high-dimensional data
    Lee, Kichun
    Gray, Alexander
    Kim, Heeyoung
    DATA MINING AND KNOWLEDGE DISCOVERY, 2013, 26 (03) : 512 - 532
  • [10] Dependence maps, a dimensionality reduction with dependence distance for high-dimensional data
    Kichun Lee
    Alexander Gray
    Heeyoung Kim
    Data Mining and Knowledge Discovery, 2013, 26 : 512 - 532