Beyond Discrete Selection: Continuous F.mbedding Space Optimization for Generative Feature Selection

被引:4
|
作者
Xiao, Meng [1 ,2 ,4 ]
Wang, Dongjie [3 ]
Wu, Min [4 ]
Wang, Pengfei [1 ,2 ]
Zhou, Yuanchun [1 ,2 ]
Fu, Yanjie [5 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Univ Cent Florida, Dept Comp Sci, Orlando, FL USA
[4] ASTAR, Inst Infocomm Res I2R, Singapore, Singapore
[5] Arizona State Univ, Sch Comp & AI, Tempe, AZ 85287 USA
来源
23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023 | 2023年
关键词
Automated Feature Selection; Continuous Space; Optimization; Deep Sequential Learning; ENSEMBLE;
D O I
10.1109/ICDM58522.2023.00078
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The goal of Feature Selection - comprising filter, wrapper, and embedded approaches - is to find the optimal feature subset for designated downstream tasks. Nevertheless, current feature selection methods are limited by: 1) the selection criteria of these methods are varied for different domains, leading them hard to he generalized; 2) the selection performance of these approaches drops significantly when processing high -dimensional feature space coupled with small sample size. In light of these challenges, we pose the question: can selected feature subsets be more robust, accurate, and input dimensionality agnostic? In this paper, we reformulate the feature selection problem as a deep differentiable optimization task and propose a new research perspective: conceptualizing discrete feature subletting as continuous embedding space optimization. We introduce a novel and principled framework that encompasses a sequential encoder, an accuracy evaluator, a sequential decoder, and a gradient ascent optimizer. This comprehensive framework inchides four important steps: preparation of features-accuracy training data, deep feature subset embedding, gradient-optimized search, and feature subset reconstruction. Specifically, we utilize reinforcement feature selection learning to generate diverse and high-quality training data and enhance generalization. By optimizing reconstruction and accuracy losses, we embed feature selection knowledge into a continuous space using an encoderevaluator-decoder model structure. We employ a gradient ascent search algorithm to find better embeddings in the learned embedding space. Furthermore, we reconstruct feature selection solutions using these embeddings and select the feature subset with the highest performance for downstream tasks as the optimal subset. Finally, extensive experimental results demonstrate the effectiveness of our proposed method, showcasing significant enhancements in feature selection robustness and accuracy. To improve the reproducibility of our research, we have released accompanying code and datasets by Dropbox.
引用
收藏
页码:688 / 697
页数:10
相关论文
共 50 条
  • [1] From Feature Selection to Continuous Optimization
    Rakhshani, Hojjat
    Idoumghar, Lhassane
    Lepagnot, Julien
    Brevilliers, Mathieu
    ARTIFICIAL EVOLUTION, EA 2019, 2020, 12052 : 1 - 12
  • [2] Efficient nonconvex sparse group feature selection via continuous and discrete optimization
    Xiang, Shuo
    Shen, Xiaotong
    Ye, Jieping
    ARTIFICIAL INTELLIGENCE, 2015, 224 : 28 - 50
  • [3] Feature selection with interactions for continuous attributes and discrete class
    Mejia-Lavalle, Manuel
    Rodriguez, Guillermo
    CERMA 2007: ELECTRONICS, ROBOTICS AND AUTOMOTIVE MECHANICS CONFERENCE, PROCEEDINGS, 2007, : 318 - +
  • [4] Discrete Biogeography Based Optimization for Feature Selection in Molecular Signatures
    Liu, Bo
    Tian, Meihong
    Zhang, Chunhua
    Li, Xiangtao
    MOLECULAR INFORMATICS, 2015, 34 (04) : 197 - 215
  • [5] Feature Selection at the Discrete Limit
    Zhang, Miao
    Ding, Chris
    Zhang, Ya
    Nie, Feiping
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 1355 - 1361
  • [6] Learning Markov Blankets for Continuous or Discrete Networks via Feature Selection
    Deng, Houtao
    Davila, Saylisse
    Runger, George
    Tuv, Eugene
    ENSEMBLES IN MACHINE LEARNING APPLICATIONS, 2011, 373 : 117 - +
  • [7] Abnormal Traffic Detection Based on Generative Adversarial Network and Feature Optimization Selection
    Ma, Wengang
    Zhang, Yadong
    Guo, Jin
    Li, Kehong
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2021, 14 (01) : 1170 - 1188
  • [8] Feature Selection Optimized by Discrete Particle Swarm Optimization for Face Recognition
    Yan, Yanjun
    Kamath, Ganapathi
    Osadciw, Lisa Ann
    OPTICS AND PHOTONICS IN GLOBAL HOMELAND SECURITY V AND BIOMETRIC TECHNOLOGY FOR HUMAN IDENTIFICATION VI, 2009, 7306
  • [9] The feature selection method for SVM with discrete particle swarm optimization algorithm
    Peng Xiyuan
    Wu Hongxing
    Peng Yu
    ISTM/2007: 7TH INTERNATIONAL SYMPOSIUM ON TEST AND MEASUREMENT, VOLS 1-7, CONFERENCE PROCEEDINGS, 2007, : 2523 - 2526
  • [10] Unsupervised Feature Selection on Networks: A Generative View
    Wei, Xiaokai
    Cao, Bokai
    Yu, Philip S.
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2215 - 2221