Dynamic selection of normalization techniques using data complexity measures

被引:134
|
作者
Jain, Sukirty [1 ]
Shukla, Sanyam [1 ]
Wadhvani, Rajesh [1 ]
机构
[1] Maulana Azad Natl Inst Technol, Bhopal 462007, Madhya Pradesh, India
关键词
Data complexity; Data preprocessing; MM-max normalization; z-score normalization; Gaussian Kernel ELM; EXTREME LEARNING-MACHINE; CLASSIFIERS; SET;
D O I
10.1016/j.eswa.2018.04.008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data preprocessing is an important step for designing classification model. Normalization is one of the preprocessing techniques used to handle the out-of-bounds attributes. This work develops 14 classification models using different learning algorithms for dynamic selection of normalization technique. This work extracts 12 data complexity measures for 48 datasets drawn from the KEEL dataset repository. Each of these datasets is normalized using min-max and z-score normalization technique. G-mean index is estimated for these normalized datasets using Gaussian Kernel Extreme Learning Machine (KELM) in order to determine the best-suited normalization technique. The data complexity measures along with the best suited normalization technique are used as an input for developing the aforementioned dynamic models. These models predict the best suitable normalization technique based on the estimated data complexity measures of the dataset The result shows that the model developed using Gaussian Kernel ELM (KELM) and Support Vector Machine (SVM) give promising results for most of the evaluated classification problems. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:252 / 262
页数:11
相关论文
共 50 条
  • [1] Using Data Complexity Measures for Thresholding in Feature Selection Rankers
    Seijo-Pardo, Borja
    Bolon-Canedo, Veronica
    Alonso-Betanzos, Amparo
    ADVANCES IN ARTIFICIAL INTELLIGENCE, CAEPIA 2016, 2016, 9868 : 121 - 131
  • [2] Data complexity measures in feature selection
    Okimoto, Lucas C.
    Lorena, Ana C.
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [3] Dynamic selection of classifiers based on complexity measures
    Schmeing, Ederson
    Brun, Andre Luiz
    Silva, Ronan Assumpcao
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 82 - 89
  • [4] Instance Ranking Using Data Complexity Measures for Training Set Selection
    Alam, Junaid
    Rani, T. Sobha
    PATTERN RECOGNITION AND MACHINE INTELLIGENCE, PREMI 2019, PT I, 2019, 11941 : 179 - 188
  • [5] Classifier selection based on data complexity measures
    Hernández-Reyes, E
    Carrasco-Ochoa, JA
    Martínez-Trinidad, JF
    PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS AND APPLICATIONS, PROCEEDINGS, 2005, 3773 : 586 - 592
  • [6] Using data complexity measures and an evolutionary cultural algorithm for gene selection in microarray data
    Sarbazi-Azad, Saeed
    Saniee Abadeh, Mohammad
    Mowlaei, Mohammad Erfan
    Saniee Abadeh, Mohammad (saniee@modares.ac.ir), 1600, Elsevier B.V. (03):
  • [7] Contribution of Data Complexity Features on Dynamic Classifier Selection
    Brun, Andre L.
    Britto, Alceu S., Jr.
    Oliveira, Luiz S.
    Enembreck, Fabricio
    Sabourin, Robert
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 4396 - 4403
  • [8] Classifier Recommendation Using Data Complexity Measures
    Garcia, Luis P. F.
    Lorena, Ana C.
    de Souto, Marcilio C. P.
    Ho, Tin Kam
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 874 - 879
  • [9] Selection of the Best Base Classifier in One-Versus-One Using Data Complexity Measures
    Moran-Fernandez, Laura
    Bolon-Canedo, Veronica
    Alonso-Betanzos, Amparo
    ADVANCES IN ARTIFICIAL INTELLIGENCE, CAEPIA 2016, 2016, 9868 : 110 - 120
  • [10] CBSN: Comparative measures of normalization techniques for brain tumor segmentation using SRCNet
    Kumar, Rahul
    Gupta, Ankur
    Arora, Harkirat Singh
    Raman, Balasubramanian
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (10) : 13203 - 13235