Efficient feature selection using shrinkage estimators

被引:24
|
作者
Sechidis, Konstantinos [1 ]
Azzimonti, Laura [2 ]
Pocock, Adam [3 ]
Corani, Giorgio [2 ]
Weatherall, James [4 ]
Brown, Gavin [1 ]
机构
[1] Univ Manchester, Sch Comp Sci, Manchester, Lancs, England
[2] Ist Dalle Molle Studi Sull Intelligenza Artificia, Manno, Switzerland
[3] Oracle Labs, Burlington, MA USA
[4] AstraZeneca, Global Med Dev, Adv Analyt Ctr, Cambridge, England
基金
英国工程与自然科学研究理事会;
关键词
Feature selection; High order feature selection; Mutual information; Shrinkage estimators; MUTUAL INFORMATION; ENTROPY; DEPENDENCIES; ALGORITHMS; INFERENCE;
D O I
10.1007/s10994-019-05795-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Information theoretic feature selection methods quantify the importance of each feature by estimating mutual information terms to capture: the relevancy, the redundancy and the complementarity. These terms are commonly estimated by maximum likelihood, while an under-explored area of research is how to use shrinkage methods instead. Our work suggests a novel shrinkage method for data-efficient estimation of information theoretic terms. The small sample behaviour makes it particularly suitable for estimation of discrete distributions with large number of categories (bins). Using our novel estimators we derive a framework for generating feature selection criteria that capture any high-order feature interaction for redundancy and complementarity. We perform a thorough empirical study across datasets from diverse sources and using various evaluation measures. Our first finding is that our shrinkage based methods achieve better results, while they keep the same computational cost as the simple maximum likelihood based methods. Furthermore, under our framework we derive efficient novel high-order criteria that outperform state-of-the-art methods in various tasks.
引用
收藏
页码:1261 / 1286
页数:26
相关论文
共 50 条
  • [31] The risk of pretest and shrinkage estimators
    Nkurunziza, Severien
    STATISTICS, 2012, 46 (03) : 305 - 312
  • [32] Shrinkage Estimators in Online Experiments
    Dimmery, Drew
    Bakshy, Eytan
    Sehkon, Jasjeet
    KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, : 2914 - 2922
  • [33] Shrinkage estimators for covariance matrices
    Daniels, MJ
    Kass, RE
    BIOMETRICS, 2001, 57 (04) : 1173 - 1184
  • [34] SHRINKAGE ESTIMATORS OF RELATIVE POTENCY
    KIM, PT
    CARTER, EM
    HUBERT, JJ
    HAND, KJ
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1993, 88 (422) : 615 - 621
  • [35] Kernel mean shrinkage estimators
    1600, Microtome Publishing (17):
  • [36] Building a Cloud IDS Using an Efficient Feature Selection Method and SVM
    Wang, Wenjuan
    Du, Xuehui
    Wang, Na
    IEEE ACCESS, 2019, 7 : 1345 - 1354
  • [37] Efficient Text Classification Using Best Feature Selection and Combination of Methods
    Srinivas, M.
    Supreethi, K. P.
    Prasad, E. V.
    Kumari, S. Anitha
    HUMAN INTERFACE AND THE MANAGEMENT OF INFORMATION: DESIGNING INFORMATION ENVIRONMENTS, PT I, 2009, 5617 : 437 - +
  • [38] An Efficient Feature Selection and Extraction using Metaheuristic Technique for Diabetic Retinopathy
    Raj, Anamika
    Noor, Noor Maizura Mohamad
    Mohemad, Rosmayati
    Mat, Noor Azliza Che
    Hussain, Shahid
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (06) : 1119 - 1131
  • [39] Efficient Feature Selection Using Weighted Superposition Attraction Optimization Algorithm
    Ganesh, Narayanan
    Shankar, Rajendran
    Cep, Robert
    Chakraborty, Shankar
    Kalita, Kanak
    APPLIED SCIENCES-BASEL, 2023, 13 (05):
  • [40] Efficient genomic selection using ensemble learning and ensemble feature reduction
    Banerjee R.
    Marathi B.
    Singh M.
    Journal of Crop Science and Biotechnology, 2020, 23 (4) : 311 - 323