Multi-label feature selection based on minimizing feature redundancy of mutual information

被引:1
|
作者
Zhou, Gaozhi [2 ]
Li, Runxin [1 ,2 ]
Shang, Zhenhong [2 ]
Li, Xiaowu [2 ]
Jia, Lianyin [2 ]
机构
[1] Kunming Univ Sci & Technol, Yunnan Key Lab Comp Technol Applicat, Kunming 650500, Peoples R China
[2] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-label feature selection; Mutual information; Sparse model; Redundant correlation; OPTIMIZATION ALGORITHM; SHRINKAGE; COMMON;
D O I
10.1016/j.neucom.2024.128392
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-label feature selection is an indispensable technology in the preprocessing of multi-label high-dimensional data. Approaches utilizing information theory and sparse models hold promise in this domain, demonstrating strong performance. Although there have been extensive literatures using l 1 and l 2 , 1-norms to identify label- specific features and common features in the feature space, they all ignore the redundant information interference problem when different features are learned simultaneously. Considering that features and labels in multi-label data are rarely linearly correlated, the MFS-MFR approach is presented to generate a representation of the nonlinear correlation between features and labels using the mutual information estimator. Following that, MFS-MFR detects specific and common features in the feature-label mutual information space using two coefficient matrices constrained by the l 1 and l 2 , 1-norms, respectively. In particular, we define a nonzero correlation constraint that effectively minimizes the redundant correlation between the two matrices. Moreover, a manifold regularization term is devised to preserve the local information of the mutual information space. To solve the optimization model with nonlinear binary regular term, we employ a novel solution approach called S-FISTA. Extensive experiments across 15 multi-label benchmark datasets, comparing against 11 top-performing multi-label feature selection methods, demonstrate the superior performance of MFS-MFR.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Multi-label feature selection based on max-dependency and min-redundancy
    Lin, Yaojin
    Hu, Qinghua
    Liu, Jinghua
    Duan, Jie
    NEUROCOMPUTING, 2015, 168 : 92 - 103
  • [42] Feature redundancy term variation for mutual information-based feature selection
    Wanfu Gao
    Liang Hu
    Ping Zhang
    Applied Intelligence, 2020, 50 : 1272 - 1288
  • [43] Multi-label learning based on instance correlation and feature redundancy
    Zhang, Yong
    Jiang, Yuqing
    Zhang, Qi
    Liu, Da
    PATTERN RECOGNITION LETTERS, 2023, 176 : 123 - 130
  • [44] Multi-label feature selection by strongly relevant label gain and label mutual aid
    Dai, Jianhua
    Huang, Weiyi
    Zhang, Chucai
    Liu, Jie
    PATTERN RECOGNITION, 2024, 145
  • [45] A Feature Selection Method for Multi-Label Text Based on Feature Importance
    Zhang, Lu
    Duan, Qingling
    APPLIED SCIENCES-BASEL, 2019, 9 (04):
  • [46] Novel multi-label feature selection via label symmetric uncertainty correlation learning and feature redundancy evaluation
    Dai, Jianhua
    Chen, Jiaolong
    Liu, Ye
    Hu, Hu
    KNOWLEDGE-BASED SYSTEMS, 2020, 207
  • [47] Partial multi-label feature selection with feature noise
    Wu, You
    Li, Peipei
    Zou, Yizhang
    PATTERN RECOGNITION, 2025, 162
  • [48] Label Construction for Multi-label Feature Selection
    Spolaor, Newton
    Monard, Maria Carolina
    Tsoumakas, Grigorios
    Lee, Huei Diana
    2014 BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 2014, : 247 - 252
  • [49] Multi-label feature selection based on correlation label enhancement
    He, Zhuoxin
    Lin, Yaojin
    Wang, Chenxi
    Guo, Lei
    Ding, Weiping
    INFORMATION SCIENCES, 2023, 647
  • [50] Multi-label feature selection based on the division of label topics
    Zhang, Ping
    Gao, Wanfu
    Hu, Juncheng
    Li, Yonghao
    INFORMATION SCIENCES, 2021, 553 : 129 - 153