Efficient heuristics for learning scalable Bayesian network classifier from labeled and unlabeled data

被引:0
|
作者
Limin Wang
Junjie Wang
Lu Guo
Qilong Li
机构
[1] Jilin University,Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education
[2] College of Software,undefined
[3] Jilin University,undefined
[4] College of Instrumentation and Electrical Engineering,undefined
[5] Jilin University,undefined
来源
Applied Intelligence | 2024年 / 54卷
关键词
Bayesian network classifier; Attribute independence assumption; Ensemble learning; Log-likelihood function; Instance learning;
D O I
暂无
中图分类号
学科分类号
摘要
Naive Bayes (NB) is one of the top ten machine learning algorithms whereas its attribute independence assumption rarely holds in practice. A feasible and efficient approach to improving NB is relaxing the assumption by adding augmented edges to the restricted topology of NB. In this paper we prove theoretically that the generalized topology may be a suboptimal solution to model multivariate probability distributions if its fitness to data cannot be measured. Thus we propose to apply log-likelihood function as the scoring function, then introduce an efficient heuristic search strategy to explore high-dependence relationships, and for each iteration the learned topology will be improved to fit data better. The proposed algorithm, called log-likelihood Bayesian classifier (LLBC), can respectively learn two submodels from labeled training set and individual unlabeled testing instance, and then make them work jointly for classification in the framework of ensemble learning. Our extensive experimental evaluations on 36 benchmark datasets from the University of California at Irvine (UCI) machine learning repository reveal that, LLBC demonstrates excellent classification performance and provides a competitive approach to learn from labeled and unlabeled data.
引用
收藏
页码:1957 / 1979
页数:22
相关论文
共 50 条
  • [41] Learning Bayesian classifiers from positive and unlabeled examples
    Calvo, Boria
    Larranaga, Pedro
    Lozano, Jose A.
    PATTERN RECOGNITION LETTERS, 2007, 28 (16) : 2375 - 2384
  • [42] A bayesian network classifier learning based on dependent analysis
    Zhang, Jianfei
    Han, Xu
    Zhang, Qi
    Wang, Shuangcheng
    ICIC Express Letters, 2013, 7 (12): : 3207 - 3212
  • [43] Estimating Accuracy from Unlabeled Data: A Bayesian Approach
    Platanios, Emmanouil Antonios
    Dubey, Avinava
    Mitchell, Tom
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [44] An Efficient Algorithm for Learning Bayesian Networks from Data
    Dojer, Norbert
    FUNDAMENTA INFORMATICAE, 2010, 103 (1-4) : 53 - 67
  • [45] Efficient learning of unlabeled term trees with contractible variables from positive data
    Suzuki, Y
    Shoudai, T
    Matsumoto, S
    Uchida, T
    INDUCTIVE LOGIC PROGRAMMING, PROCEEDINGS, 2003, 2835 : 347 - 364
  • [46] Discriminative clustering with representation learning with any ratio of labeled to unlabeled data
    Jones, Corinne
    Roulet, Vincent
    Harchaoui, Zaid
    STATISTICS AND COMPUTING, 2022, 32 (01)
  • [47] Discriminative clustering with representation learning with any ratio of labeled to unlabeled data
    Corinne Jones
    Vincent Roulet
    Zaid Harchaoui
    Statistics and Computing, 2022, 32
  • [48] A survey on Bayesian network structure learning from data
    Mauro Scanagatta
    Antonio Salmerón
    Fabio Stella
    Progress in Artificial Intelligence, 2019, 8 : 425 - 439
  • [49] LEARNING BAYESIAN NETWORK PARAMETERS FROM SOFT DATA
    Xiao, Xu Hong
    Lee, Hian Beng
    Ng, Gee Wah
    INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2009, 17 (02) : 281 - 294
  • [50] The Use of Unlabeled Data versus Labeled Data for Stopping Active Learning for Text Classification
    Beatty, Garrett
    Kochis, Ethan
    Bloodgood, Michael
    2019 13TH IEEE INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC), 2019, : 287 - 294