Robust feature selection via l2,1-norm in finite mixture of regression

被引:1
|
作者
Li, Xiangrui [1 ]
Zhu, Dongxiao [1 ]
机构
[1] Wayne State Univ, Dept Comp Sci, Detroit, MI 48202 USA
基金
美国国家科学基金会;
关键词
Finite mixture of regression; Feature selection; Non-convex optimization; HIERARCHICAL MIXTURES; CLASSIFICATION; EXPERTS; MODEL;
D O I
10.1016/j.patrec.2018.02.021
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Finite mixture of Gaussian regression (FMR) is a widely-used modeling technique in supervised learning problems. In cases where the number of features is large, feature selection is desirable to enhance model interpretability and to avoid overfitting. In this paper, we propose a robust feature selection method via l(2,1)-norm penalized maximum likelihood estimation (MLE) in FMR, with extension to sparse l(2,1) penalty by combining l(1)-norm with l(2,1)-norm for increasing flexibility. To solve the non-convex and non-smooth problem of (sparse) penalized MLE in FMR, we develop an new EM-based algorithm for numerical optimization, with combination of block coordinate descent and majorizing-mimmization scheme in M-step. We finally apply our method in six simulations and one real dataset to demonstrate its superior performance. (c) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:15 / 22
页数:8
相关论文
共 50 条
  • [1] Cost-sensitive feature selection via the l2,1-norm
    Zhao, Hong
    Yu, Shenglong
    [J]. INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2019, 104 : 25 - 37
  • [2] Robust discriminant feature selection via joint L2,1-norm distance minimization and maximization
    Yang, Zhangjing
    Ye, Qiaolin
    Chen, Qiao
    Ma, Xu
    Fu, Liyong
    Yang, Guowei
    Yan, He
    Liu, Fan
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 207
  • [3] Unsupervised maximum margin feature selection via L2,1-norm minimization
    Shizhun Yang
    Chenping Hou
    Feiping Nie
    Yi Wu
    [J]. Neural Computing and Applications, 2012, 21 : 1791 - 1799
  • [4] l2,1-norm based Regression for Classification
    Ren, Chuan-Xian
    Dai, Dao-Qing
    Yan, Hong
    [J]. 2011 FIRST ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2011, : 485 - 489
  • [5] Unsupervised Discriminative Feature Selection in a Kernel Space via L2,1-Norm Minimization
    Liu, Yang
    Wang, Yizhou
    [J]. 2012 21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR 2012), 2012, : 1205 - 1208
  • [6] Robust classification using l2,1-norm based regression model
    Ren, Chuan-Xian
    Dai, Dao-Qing
    Yan, Hong
    [J]. PATTERN RECOGNITION, 2012, 45 (07) : 2708 - 2718
  • [7] l2,1-norm minimization based negative label relaxation linear regression for feature selection
    Peng, Yali
    Sehdev, Paramjit
    Liu, Shigang
    Lie, Jun
    Wang, Xili
    [J]. PATTERN RECOGNITION LETTERS, 2018, 116 : 170 - 178
  • [8] Discriminative Feature Selection via Joint Trace Ratio Criterion and l2,1-norm Regularization
    Jiang, Zhang
    Zhao, Mingbo
    Kong, Weijian
    [J]. 2018 IEEE SYMPOSIUM ON PRODUCT COMPLIANCE ENGINEERING - ASIA 2018 (IEEE ISPCE-CN 2018), 2018, : 27 - 32
  • [9] Robust Feature Selection via Simultaneous Capped l2 -Norm and l2,1 -Norm Minimization
    Lan, Gongmin
    Hou, Chenping
    Yi, Dongyun
    [J]. PROCEEDINGS OF 2016 IEEE INTERNATIONAL CONFERENCE ON BIG DATA ANALYSIS (ICBDA), 2016, : 147 - 151
  • [10] Fast unsupervised feature selection with anchor graph and l2,1-norm regularization
    Hu, Haojie
    Wang, Rong
    Nie, Feiping
    Yang, Xiaojun
    Yu, Weizhong
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (17) : 22099 - 22113