Subquadratic High-Dimensional Hierarchical Clustering

被引:0
|
作者
Abboud, Amir [1 ]
Cohen-Addad, Vincent [2 ,3 ]
Houdrouge, Hussein [4 ]
机构
[1] IBM Res, Yorktown Hts, NY 10598 USA
[2] CNRS, Paris, France
[3] Sorbonne Univ, Paris, France
[4] Ecole Polytech, Palaiseau, France
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the widely-used average-linkage, single-linkage, and Ward's methods for computing hierarchical clusterings of high-dimensional Euclidean inputs. It is easy to show that there is no efficient implementation of these algorithms in high dimensional Euclidean space since it implicitly requires to solve the closest pair problem, a notoriously difficult problem. However, how fast can these algorithms be implemented if we allow approximation? More precisely: these algorithms successively merge the clusters that are at closest average (for average-linkage), minimum distance (for single-linkage), or inducing the least sum-of-square error (for Ward's). We ask whether one could obtain a significant running-time improvement if the algorithm can merge a-approximate closest clusters (namely, clusters that are at distance (average, minimum, or sum-of-square error) at most gamma times the distance of the closest clusters). We show that one can indeed take advantage of the relaxation and compute the approximate hierarchical clustering tree using (O) over tilde (n) gamma-approximate nearest neighbor queries. This leads to an algorithm running in time (O) over tilde (nd) + n(1+O(1/gamma)) for d-dimensional Euclidean space. We then provide experiments showing that these algorithms perform as well as the non-approximate version for classic classification tasks while achieving a significant speed-up.
引用
收藏
页数:11
相关论文
共 50 条