General Cost Models for Evaluating Dimensionality Reduction in High-Dimensional Spaces

被引:8
|
作者
Lian, Xiang [1 ]
Chen, Lei [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Kowloon, Hong Kong, Peoples R China
关键词
High-dimensionality reduction; similarity search;
D O I
10.1109/TKDE.2008.170
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Similarity search usually encounters a serious problem in the high-dimensional space, known as the "curse of dimensionality." In order to speed up the retrieval efficiency, most previous approaches reduce the dimensionality of the entire data set to a fixed lower value before building indexes (referred to as global dimensionality reduction (GDR)). More recent works focus on locally reducing the dimensionality of data to different values (called the local dimensionality reduction (LDR)). In addition, random projection is proposed as an approximate dimensionality reduction (ADR) technique to answer the approximate similarity search instead of the exact one. However, so far little work has formally evaluated the effectiveness and efficiency of GDR, LDR, and ADR for the range query. Motivated by this, in this paper, we propose general cost models for evaluating the query performance over the reduced data sets by GDR, LDR, and ADR, in light of which we introduce a novel (A) LDR method, Partitioning based on RANdomized Search (PRANS). It can achieve high retrieval efficiency with the guarantee of optimality given by the formal models. Finally, a B(+)-tree index is constructed over the reduced partitions for fast similarity search. Extensive experiments validate the correctness of our cost models on both real and synthetic data sets and demonstrate the efficiency and effectiveness of the proposed PRANS method.
引用
收藏
页码:1447 / 1460
页数:14
相关论文
共 50 条
  • [31] Posterior Consistency of Factor Dimensionality in High-Dimensional Sparse Factor Models
    Ohn, Ilsang
    Kim, Yongdai
    [J]. BAYESIAN ANALYSIS, 2022, 17 (02): : 491 - 514
  • [32] A comprehensive empirical comparison of hubness reduction in high-dimensional spaces
    Feldbauer, Roman
    Flexer, Arthur
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2019, 59 (01) : 137 - 166
  • [33] A comprehensive empirical comparison of hubness reduction in high-dimensional spaces
    Roman Feldbauer
    Arthur Flexer
    [J]. Knowledge and Information Systems, 2019, 59 : 137 - 166
  • [34] A sparse grid based method for generative dimensionality reduction of high-dimensional data
    Bohn, Bastian
    Garcke, Jochen
    Griebel, Michael
    [J]. JOURNAL OF COMPUTATIONAL PHYSICS, 2016, 309 : 1 - 17
  • [35] Dimension Reduction for High-Dimensional Vector Autoregressive Models
    Cubadda, Gianluca
    Hecq, Alain
    [J]. OXFORD BULLETIN OF ECONOMICS AND STATISTICS, 2022, 84 (05) : 1123 - 1152
  • [36] Glyphboard: Visual Exploration of High-Dimensional Data Combining Glyphs with Dimensionality Reduction
    Kammer, Dietrich
    Keck, Mandy
    Gruender, Thomas
    Maasch, Alexander
    Thom, Thomas
    Kleinsteuber, Martin
    Groh, Rainer
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (04) : 1661 - 1671
  • [37] Hierarchical surrogate model with dimensionality reduction technique for high-dimensional uncertainty propagation
    Cheng, Kai
    Lu, Zhenzhou
    [J]. INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN ENGINEERING, 2020, 121 (09) : 2068 - 2085
  • [38] Rank reduction for high-dimensional generalized additive models
    Lin, Hongmei
    Lian, Heng
    Liang, Hua
    [J]. JOURNAL OF MULTIVARIATE ANALYSIS, 2019, 173 : 672 - 684
  • [39] Recent Dimensionality Reduction Techniques for High-Dimensional COVID-19 Data
    Dallas, Ioannis L.
    Vrahatis, Aristidis G.
    Tasoulis, Sotiris K.
    Plagianakos, Vassilis P.
    [J]. COMPUTATIONAL INTELLIGENCE METHODS FOR BIOINFORMATICS AND BIOSTATISTICS, CIBB 2021, 2022, 13483 : 227 - 241
  • [40] Distance-preserving projection of high-dimensional data for nonlinear dimensionality reduction
    Yang, L
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2004, 26 (09) : 1243 - 1246