Simultaneous feature selection and clustering using mixture models

被引:429
|
作者
Law, MHC
Figueiredo, MAT
Jain, AK
机构
[1] Michigan State Univ, Dept Comp Sci & Engn, E Lansing, MI 48824 USA
[2] Inst Super Tecn, Inst Telecomunicacoes, P-1049001 Lisbon, Portugal
关键词
feature selection; clustering; unsupervised learning; mixture models; minimum message length; EM algorithm;
D O I
10.1109/TPAMI.2004.71
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Clustering is a common unsupervised learning technique used to discover group structure in a set of data. While there exist many algorithms for clustering, the important issue of feature selection, that is, what attributes of the data should be used by the clustering algorithms, is rarely touched upon. Feature selection for clustering is difficult because, unlike in supervised learning, there are no class labels for the data and, thus, no obvious criteria to guide the search. Another important problem in clustering is the determination of the number of clusters, which clearly impacts and is influenced by the feature selection issue. In this paper, we propose the concept of feature saliency and introduce an expectation-maximization (EM) algorithm to estimate it, in the context of mixture-based clustering. Due to the introduction of a minimum message length model selection criterion, the saliency of irrelevant features is driven toward zero, which corresponds to performing feature selection. The criterion and algorithm are then extended to simultaneously estimate the feature saliencies and the number of clusters.
引用
下载
收藏
页码:1154 / 1166
页数:13
相关论文
共 50 条