Efficient feature selection using shrinkage estimators

被引:0
|
作者
Konstantinos Sechidis
Laura Azzimonti
Adam Pocock
Giorgio Corani
James Weatherall
Gavin Brown
机构
[1] University of Manchester,School of Computer Science
[2] Istituto Dalle Molle di studi sull’ Intelligenza Artificiale (IDSIA),Advanced Analytics Centre, Global Medicines Development
[3] Oracle Labs,undefined
[4] AstraZeneca,undefined
来源
Machine Learning | 2019年 / 108卷
关键词
Feature selection; High order feature selection; Mutual information; Shrinkage estimators;
D O I
暂无
中图分类号
学科分类号
摘要
Information theoretic feature selection methods quantify the importance of each feature by estimating mutual information terms to capture: the relevancy, the redundancy and the complementarity. These terms are commonly estimated by maximum likelihood, while an under-explored area of research is how to use shrinkage methods instead. Our work suggests a novel shrinkage method for data-efficient estimation of information theoretic terms. The small sample behaviour makes it particularly suitable for estimation of discrete distributions with large number of categories (bins). Using our novel estimators we derive a framework for generating feature selection criteria that capture any high-order feature interaction for redundancy and complementarity. We perform a thorough empirical study across datasets from diverse sources and using various evaluation measures. Our first finding is that our shrinkage based methods achieve better results, while they keep the same computational cost as the simple maximum likelihood based methods. Furthermore, under our framework we derive efficient novel high-order criteria that outperform state-of-the-art methods in various tasks.
引用
收藏
页码:1261 / 1286
页数:25
相关论文
共 50 条
  • [1] Efficient feature selection using shrinkage estimators
    Sechidis, Konstantinos
    Azzimonti, Laura
    Pocock, Adam
    Corani, Giorgio
    Weatherall, James
    Brown, Gavin
    [J]. MACHINE LEARNING, 2019, 108 (8-9) : 1261 - 1286
  • [2] Correction to: Efficient feature selection using shrinkage estimators
    Konstantinos Sechidis
    Laura Azzimonti
    Adam Pocock
    Giorgio Corani
    James Weatherall
    Gavin Brown
    [J]. Machine Learning, 2020, 109 : 1565 - 1567
  • [3] Efficient feature selection using shrinkage estimators (vol 108, pg 1261, 2019)
    Sechidis, Konstantinos
    Azzimonti, Laura
    Pocock, Adam
    Corani, Giorgio
    Weatherall, James
    Brown, Gavin
    [J]. MACHINE LEARNING, 2020, 109 (08) : 1565 - 1567
  • [4] Feature selection with missing data using mutual information estimators
    Doquire, Gauthier
    Verleysen, Michel
    [J]. NEUROCOMPUTING, 2012, 90 : 3 - 11
  • [5] Feature selection and feature learning in arousal dimension of music emotion by using shrinkage methods
    Jiang Long Zhang
    Xiang Lin Huang
    Li Fang Yang
    Ye Xu
    Shu Tao Sun
    [J]. Multimedia Systems, 2017, 23 : 251 - 264
  • [6] Feature selection and feature learning in arousal dimension of music emotion by using shrinkage methods
    Zhang, Jiang Long
    Huang, Xiang Lin
    Yang, Li Fang
    Xu, Ye
    Sun, Shu Tao
    [J]. MULTIMEDIA SYSTEMS, 2017, 23 (02) : 251 - 264
  • [7] Size matters: Optimal calibration of shrinkage estimators for portfolio selection
    DeMiguel, Victor
    Martin-Utrera, Alberto
    Nogales, Francisco J.
    [J]. JOURNAL OF BANKING & FINANCE, 2013, 37 (08) : 3018 - 3034
  • [8] Efficient Estimation and Validation of Shrinkage Estimators in Big Data Analytics
    du Plessis, Salomi
    Arashi, Mohammad
    Maribe, Gaonyalelwe
    Millard, Salomon M.
    [J]. MATHEMATICS, 2023, 11 (22)
  • [9] Efficient Feature Ranking and Selection Using Statistical Moments
    Hochma, Yael
    Felendler, Yuval
    Last, Mark
    [J]. IEEE ACCESS, 2024, 12 : 105573 - 105587
  • [10] Efficient Feature Selection for Prediction of Diabetic Using LASSO
    Kumarage, Prabha M.
    Yogarajah, B.
    Ratnarajah, Nagulan
    [J]. 2019 19TH INTERNATIONAL CONFERENCE ON ADVANCES IN ICT FOR EMERGING REGIONS (ICTER - 2019), 2019,