Feature redundancy term variation for mutual information-based feature selection

被引:39
|
作者
Gao, Wanfu [1 ,2 ,3 ]
Hu, Liang [1 ,2 ]
Zhang, Ping [1 ,2 ]
机构
[1] JiLin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, Minist Educ, Changchun, Peoples R China
[3] Jilin Univ, Coll Chem, Changchun, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Machine learning; Feature selection; Information theory; Feature redundancy; RELEVANCE;
D O I
10.1007/s10489-019-01597-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection plays a critical role in many applications that are relevant to machine learning, image processing and gene expression analysis. Traditional feature selection methods intend to maximize feature dependency while minimizing feature redundancy. In previous information-theoretical-based feature selection methods, feature redundancy term is measured by the mutual information between a candidate feature and each already-selected feature or the interaction information among a candidate feature, each already-selected feature and the class. However, the larger values of the traditional feature redundancy term do not indicate the worse a candidate feature because a candidate feature can obtain large redundant information, meanwhile offering large new classification information. To address this issue, we design a new feature redundancy term that considers the relevancy between a candidate feature and the class given each already-selected feature, and a novel feature selection method named min-redundancy and max-dependency (MRMD) is proposed. To verify the effectiveness of our method, MRMD is compared to eight competitive methods on an artificial example and fifteen real-world data sets respectively. The experimental results show that our method achieves the best classification performance with respect to multiple evaluation criteria.
引用
收藏
页码:1272 / 1288
页数:17
相关论文
共 50 条
  • [1] Feature redundancy term variation for mutual information-based feature selection
    Wanfu Gao
    Liang Hu
    Ping Zhang
    [J]. Applied Intelligence, 2020, 50 : 1272 - 1288
  • [2] Conditional Mutual Information-Based Feature Selection Analyzing for Synergy and Redundancy
    Cheng, Hongrong
    Qin, Zhiguang
    Feng, Chaosheng
    Wang, Yong
    Li, Fagen
    [J]. ETRI JOURNAL, 2011, 33 (02) : 210 - 218
  • [3] Mutual information-based feature selection for radiomics
    Oubel, Estanislao
    Beaumont, Hubert
    Iannessi, Antoine
    [J]. MEDICAL IMAGING 2016: PACS AND IMAGING INFORMATICS: NEXT GENERATION AND INNOVATIONS, 2016, 9789
  • [4] Conditional mutual information-based feature selection algorithm for maximal relevance minimal redundancy
    Gu, Xiangyuan
    Guo, Jichang
    Xiao, Lijun
    Li, Chongyi
    [J]. APPLIED INTELLIGENCE, 2022, 52 (02) : 1436 - 1447
  • [5] Conditional mutual information-based feature selection algorithm for maximal relevance minimal redundancy
    Xiangyuan Gu
    Jichang Guo
    Lijun Xiao
    Chongyi Li
    [J]. Applied Intelligence, 2022, 52 : 1436 - 1447
  • [6] Mutual information-based feature selection for multilabel classification
    Doquire, Gauthier
    Verleysen, Michel
    [J]. NEUROCOMPUTING, 2013, 122 : 148 - 155
  • [7] Stopping rules for mutual information-based feature selection
    Mielniczuk, Jan
    Teisseyre, Pawel
    [J]. NEUROCOMPUTING, 2019, 358 : 255 - 274
  • [8] A Study on Mutual Information-Based Feature Selection in Classifiers
    Arundhathi, B.
    Athira, A.
    Rajan, Ranjidha
    [J]. ARTIFICIAL INTELLIGENCE AND EVOLUTIONARY COMPUTATIONS IN ENGINEERING SYSTEMS, ICAIECES 2016, 2017, 517 : 479 - 486
  • [9] CONDITIONAL DYNAMIC MUTUAL INFORMATION-BASED FEATURE SELECTION
    Liu, Huawen
    Mo, Yuchang
    Zhao, Jianmin
    [J]. COMPUTING AND INFORMATICS, 2012, 31 (06) : 1193 - 1216
  • [10] Heterogeneous feature subset selection using mutual information-based feature transformation
    Wei, Min
    Chow, Tommy W. S.
    Chan, Rosa H. M.
    [J]. NEUROCOMPUTING, 2015, 168 : 706 - 718