Bayes Theorem and Information Gain Based Feature Selection for Maximizing the Performance of Classifiers

被引:0
|
作者
Appavu, Subramanian
Rajaram, Ramasamy
Nagammai, M.
Priyanga, N.
Priyanka, S.
机构
关键词
Data mining; Feature Selection; Classification; Bayes Theorem and Information Gain; DECISION TREE INDUCTION;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Features play a very important role in the task of pattern classification. Consequently, the selection of suitable features is necessary as most of the raw data might be redundant or irrelevant to the recognition of patterns. In some cases, the classifier can not perform well because of the large number of redundant features. This paper presents a novel evolving feature selection algorithms taking the advantages of Bayes Theorem and Information Gain to improve the predictive accuracy. Bayes theorem is used to discover dependency information among features. In addition to that, feature selection has been improvised by Information Gain which selects features based on their importance. Different features play different roles in classifying datasets. Unwanted features will result in error information during classification which will reduce classification precision. The proposed feature selection can remove these distractions to improve classification performance. As shown in the experimental results, after feature selection using the Bayes theorem and Information gain to control false discovery rate, the classification performance of DT's and NB classifiers were significantly improved.
引用
收藏
页码:501 / 511
页数:11
相关论文
共 50 条
  • [1] An unsupervised feature selection algorithm with feature ranking for maximizing performance of the classifiers
    Singh D.A.A.G.
    Balamurugan S.A.A.
    Leavline E.J.
    [J]. International Journal of Automation and Computing, 2015, 12 (05) : 511 - 517
  • [2] An Unsupervised Feature Selection Algorithm with Feature Ranking for Maximizing Performance of the Classifiers
    Danasingh Asir Antony Gnana Singh
    Subramanian Appavu Alias Balamurugan
    Epiphany Jebamalar Leavline
    [J]. Machine Intelligence Research, 2015, (05) : 511 - 517
  • [3] Feature selection via maximizing global information gain for text classification
    Shang, Changxing
    Li, Min
    Feng, Shengzhong
    Jiang, Qingshan
    Fan, Jianping
    [J]. KNOWLEDGE-BASED SYSTEMS, 2013, 54 : 298 - 309
  • [4] Iterative Feature Selection using Information Gain & Naive Bayes for Document Classification
    Rahman, Chowdhury Mofizur
    Afroze, Lameya
    Refath, Naznin Sultana
    Shawon, Nafin
    [J]. 2018 21ST INTERNATIONAL CONFERENCE OF COMPUTER AND INFORMATION TECHNOLOGY (ICCIT), 2018,
  • [5] Gabor Feature Selection Based on Information Gain
    Lefkovits, Szidonia
    Lefkovits, Laszlo
    [J]. 10TH INTERNATIONAL CONFERENCE INTERDISCIPLINARITY IN ENGINEERING, INTER-ENG 2016, 2017, 181 : 892 - 898
  • [6] A New Feature Selection Approach to Naive Bayes Text Classifiers
    Zhang, Lungan
    Jiang, Liangxiao
    Li, Chaoqun
    [J]. INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2016, 30 (02)
  • [7] A Study on Mutual Information-Based Feature Selection in Classifiers
    Arundhathi, B.
    Athira, A.
    Rajan, Ranjidha
    [J]. ARTIFICIAL INTELLIGENCE AND EVOLUTIONARY COMPUTATIONS IN ENGINEERING SYSTEMS, ICAIECES 2016, 2017, 517 : 479 - 486
  • [8] Optimized Approach of Feature Selection based on Information Gain
    Wu, Guohua
    Xu, Junjun
    [J]. 2015 INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND MECHANICAL AUTOMATION (CSMA), 2015, : 157 - 161
  • [9] An Improved Feature Selection Method Based on Information Gain
    Li, Yanling
    Sun, Wenxia
    [J]. INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING BIOMEDICAL ENGINEERING, AND INFORMATICS (SPBEI 2013), 2014, : 530 - 535
  • [10] Feature Selection by Maximizing Independent Classification Information
    Wang, Jun
    Wei, Jin-Mao
    Yang, Zhenglu
    Wang, Shu-Qin
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2017, 29 (04) : 828 - 841