Feature Selection With Redundancy-Constrained Class Separability

被引:46
|
作者
Zhou, Luping [1 ]
Wang, Lei [1 ]
Shen, Chunhua [2 ]
机构
[1] Australian Natl Univ, Sch Engn, Canberra, ACT 0200, Australia
[2] NICTA, Canberra Res Lab, Canberra, ACT 2601, Australia
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2010年 / 21卷 / 05期
基金
澳大利亚研究理事会;
关键词
Class separability measure; feature redundancy; feature selection; fractional programming; integer programming; CLASSIFICATION; ALGORITHMS;
D O I
10.1109/TNN.2010.2044189
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scatter-matrix-based class separability is a simple and efficient feature selection criterion in the literature. However, the conventional trace-based formulation does not take feature redundancy into account and is prone to selecting a set of discriminative but mutually redundant features. In this brief, we first theoretically prove that in the context of this trace-based criterion the existence of sufficiently correlated features can always prevent selecting the optimal feature set. Then, on top of this criterion, we propose the redundancy-constrained feature selection (RCFS). To ensure the algorithm's efficiency and scalability, we study the characteristic of the constraints with which the resulted constrained 0-1 optimization can be efficiently and globally solved. By using the totally unimodular (TUM) concept in integer programming, a necessary condition for such constraints is derived. This condition reveals an interesting special case in which qualified redundancy constraints can be conveniently generated via a clustering of features. We study this special case and develop an efficient feature selection approach based on Dinkelbach's algorithm. Experiments on benchmark data sets demonstrate the superior performance of our approach to those without redundancy constraints.
引用
收藏
页码:853 / 858
页数:6
相关论文
共 50 条
  • [31] Feature selection via maximizing inter-class independence and minimizing intra-class redundancy for hierarchical classification
    Shi, Jie
    Li, Zhengyu
    Zhao, Hong
    [J]. INFORMATION SCIENCES, 2023, 626 : 1 - 18
  • [32] IPIC SEPARABILITY RATIO FOR SEMI-SUPERVISED FEATURE SELECTION
    Yeung, Daniel S.
    Wang, Jun
    Ng, Wing W. Y.
    [J]. PROCEEDINGS OF 2009 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-6, 2009, : 399 - +
  • [33] Simple generative model for assessing feature selection based on relevance, redundancy, and redundancy
    Theiler, James
    [J]. APPLICATIONS OF MACHINE LEARNING, 2019, 11139
  • [34] Class-specific feature selection using neighborhood mutual information with relevance-redundancy weight
    Ma, Xi-Ao
    Lu, Kecheng
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [35] An Empirical Evaluation of Constrained Feature Selection
    Bach J.
    Zoller K.
    Trittenbach H.
    Schulz K.
    Böhm K.
    [J]. SN Computer Science, 3 (6)
  • [36] Constrained Feature Selection for Localizing Faults
    Le, Tien-Duy B.
    Lo, David
    Li, Ming
    [J]. 2015 31ST INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME) PROCEEDINGS, 2015, : 501 - 505
  • [37] Fast-Ensembles of Minimum Redundancy Feature Selection
    Schowe, Benjamin
    Morik, Katharina
    [J]. ENSEMBLES IN MACHINE LEARNING APPLICATIONS, 2011, 373 : 75 - 95
  • [38] A Feature Selection Method Based on New Redundancy Measurement
    Li, Zhan-Shan
    Lyu, Ai-Na
    [J]. Dongbei Daxue Xuebao/Journal of Northeastern University, 2020, 41 (11): : 1550 - 1556
  • [39] Feature Selection with Ensembles, Artificial Variables, and Redundancy Elimination
    Tuv, Eugene
    Borisov, Alexander
    Runger, George
    Torkkola, Kari
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2009, 10 : 1341 - 1366
  • [40] Weighted structure preservation and redundancy minimization for feature selection
    Ye, Qing
    Sun, Yaxin
    [J]. SOFT COMPUTING, 2018, 22 (21) : 7255 - 7268