Information-theoretic partially labeled heterogeneous feature selection based on neighborhood rough sets

被引:14
|
作者
Zhang, Hongying [1 ]
Sun, Qianqian [1 ]
Dong, Kezhen [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature selection; Monotonic entropy; Partially labeled heterogeneous data; ATTRIBUTE REDUCTION;
D O I
10.1016/j.ijar.2022.12.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid increase of large-scale, real-world datasets, it becomes critical to address the problem of partially labeled heterogeneous feature selection (i.e., some samples, which own numerical and categorical features, have no labels). Existing solutions typically adopt linear correlations between features. In this paper, three different monotonic uncertainty measures are defined on equivalence classes and neighborhood classes to study the partially labeled heterogeneous feature selection by exploring the nonlinear correlations. First, consistent entropy and monotonic neighborhood entropy, based on classical rough set theory and neighborhood rough set theory, are proposed to construct a uniform measure for feature selection in heterogeneous datasets. Furthermore, a maximal neighborhood entropy strategy is developed by considering the inconsistency of neighborhood classes described by the features and partial labels. Finally, two feature selection algorithms are presented by three novel monotonic uncertainty measures. The comparative experiments demonstrate the effectiveness and superiority of the newly proposed feature selection measures.(c) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:200 / 217
页数:18
相关论文
共 50 条
  • [41] Information-Theoretic Feature Selection via Tensor Decomposition and Submodularity
    Amiridi, Magda
    Kargas, Nikos
    Sidiropoulos, Nicholas D.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 6195 - 6205
  • [42] A Rigorous Information-Theoretic Definition of Redundancy and Relevancy in Feature Selection Based on (Partial) Information Decomposition
    Wollstadt, Patricia
    Schmitt, Sebastian
    Wibral, Michael
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [43] An improved runner-root algorithm for solving feature selection problems based on rough sets and neighborhood rough sets
    Ibrahim, Rehab Ali
    Abd Elaziz, Mohamed
    Oliva, Diego
    Lu, Songfeng
    APPLIED SOFT COMPUTING, 2020, 97
  • [44] Feature subset selection wrapper based on mutual information and rough sets
    Foithong, Sombut
    Pinngern, Ouen
    Attachoo, Boonwat
    EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (01) : 574 - 584
  • [45] A novel adaptive neighborhood rough sets based on sparrow search algorithm and feature selection
    Liu, Caihui
    Lin, Bowen
    Miao, Duoqian
    INFORMATION SCIENCES, 2024, 679
  • [46] Feature Weighting and Instance Selection for Collaborative Filtering: An Information-Theoretic Approach*
    Kai Yu
    Xiaowei Xu
    Martin Ester
    Hans-Peter Kriegel
    Knowledge and Information Systems, 2003, 5 (2) : 201 - 224
  • [47] Feature selection for blind image steganalysis using neighborhood rough sets
    Chen, Yingyue
    Chen, Yumin
    Yin, Aimin
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 37 (03) : 3709 - 3720
  • [48] Information-theoretic approaches to SVM feature selection for metagenome read classification
    Garbarine, Elaine
    DePasquale, Joseph
    Gadia, Vinay
    Polikar, Robi
    Rosen, Gail
    COMPUTATIONAL BIOLOGY AND CHEMISTRY, 2011, 35 (03) : 199 - 209
  • [49] Generalized Information-Theoretic Criterion for Multi-Label Feature Selection
    Seo, Wangduk
    Kim, Dae-Won
    Lee, Jaesung
    IEEE ACCESS, 2019, 7 (122854-122863) : 122854 - 122863
  • [50] Information-Theoretic Feature Selection in Microarray Data Using Variable Complementarity
    Meyer, Patrick Emmanuel
    Schretter, Colas
    Bontempi, Gianluca
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2008, 2 (03) : 261 - 274