A boosting framework for positive-unlabeled learning

被引:0
|
作者
Zhao, Yawen [1 ]
Zhang, Mingzhe [1 ]
Zhang, Chenhao [1 ]
Chen, Weitong [2 ]
Ye, Nan [1 ]
Xu, Miao [1 ]
机构
[1] Univ Queensland, Brisbane, Qld, Australia
[2] Univ Adelaide, Adelaide, SA, Australia
基金
澳大利亚研究理事会;
关键词
Boosting; Weakly supervised learning; PU learning; Ensemble;
D O I
10.1007/s11222-024-10529-y
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Positive-unlabeled (PU) learning deals with binary classification problems where only positive and unlabeled data are available. In this paper, we introduce a novel boosting framework, Adaptive PU (AdaPU), for learning from PU data. AdaPU builds an ensemble of weak classifiers using weak learners tailored to PU data. We propose two main approaches for learning the weak classifiers: a direct loss minimization approach that learns weak classifiers to greedily minimize PU-data-based estimates of the exponential loss, specifically, the unbiased PU estimate and the non-negative PU estimate; and a constrained loss minimization approach that learns weak classifiers to greedily minimize the unbiased PU estimate of the exponential loss, subject to regularization constraints. The direct loss minimization approach, while natural and simple, often yields weak learners prone to overfitting or leads to computationally expensive algorithms. On the other hand, the constrained loss minimization approach can effectively alleviate overfitting and allow the design of efficient weak learners. In particular, we propose a tailored weak learner for the simple class of decision stumps, or one-level decision trees, which interestingly demonstrates strong performance in comparison to various other weak classifiers. Furthermore, we provide several theoretical results on the performance of AdaPU. We performed extensive experiments to evaluate the variants of AdaPU and various baseline algorithms. Our results demonstrate the effectiveness of the constrained loss minimization approach.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Density Estimators for Positive-Unlabeled Learning
    Basile, Teresa M. A.
    Di Mauro, Nicola
    Esposito, Floriana
    Ferilli, Stefano
    Vergari, Antonio
    NEW FRONTIERS IN MINING COMPLEX PATTERNS, NFMCP 2017, 2018, 10785 : 49 - 64
  • [2] Generative Adversarial Positive-Unlabeled Learning
    Hou, Ming
    Chaib-draa, Brahim
    Li, Chao
    Zhao, Qibin
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2255 - 2261
  • [3] Positive-Unlabeled Learning in Streaming Networks
    Chang, Shiyu
    Zhang, Yang
    Tang, Jiliang
    Yin, Dawei
    Chang, Yi
    Hasegawa-Johnson, Mark A.
    Huang, Thomas S.
    KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 755 - 764
  • [4] Positive-Unlabeled Learning for Knowledge Distillation
    Ning Jiang
    Jialiang Tang
    Wenxin Yu
    Neural Processing Letters, 2023, 55 : 2613 - 2631
  • [5] An Integrated Framework of Positive-Unlabeled and Imbalanced Learning for Landslide Susceptibility Mapping
    Fu, Zijin
    Ma, Hao
    Wang, Fawu
    Dou, Jie
    Zhang, Bo
    Fang, Zhice
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 15596 - 15611
  • [6] Positive-Unlabeled Learning for Knowledge Distillation
    Jiang, Ning
    Tang, Jialiang
    Yu, Wenxin
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2613 - 2631
  • [7] Positive-Unlabeled Learning With Label Distribution Alignment
    Jiang, Yangbangyan
    Xu, Qianqian
    Zhao, Yunrui
    Yang, Zhiyong
    Wen, Peisong
    Cao, Xiaochun
    Huang, Qingming
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 15345 - 15363
  • [8] Positive-Unlabeled Learning for Network Link Prediction
    Gan, Shengfeng
    Alshahrani, Mohammed
    Liu, Shichao
    MATHEMATICS, 2022, 10 (18)
  • [9] Positive-Unlabeled Learning from Imbalanced Data
    Su, Guangxin
    Chen, Weitong
    Xu, Miao
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2995 - 3001
  • [10] Ensembles of density estimators for positive-unlabeled learning
    T. M. A. Basile
    N. Di Mauro
    F. Esposito
    S. Ferilli
    A. Vergari
    Journal of Intelligent Information Systems, 2019, 53 : 199 - 217