Semi-supervised trees for multi-target regression

被引:36
|
作者
Levatic, Jurica [1 ,2 ]
Kocev, Dragi [1 ,2 ,3 ]
Ceci, Michelangelo [3 ,4 ]
Dzeroski, Saso [1 ,2 ]
机构
[1] Josef Stefan Inst, Dept Knowledge Technol, Ljubljana, Slovenia
[2] Jotef Stefan Int Postgrad Sch, Ljubljana, Slovenia
[3] Univ Bari Aldo Moro, Dept Comp Sci, Bari, Italy
[4] CINI, Rome, Italy
基金
欧盟地平线“2020”;
关键词
Semi-supervised learning; Multi-target regression; Structured outputs; Predictive clustering trees; Random forests; CLASSIFICATION; MODEL; PREDICTION; INDUCTION; ENSEMBLES; INDEX;
D O I
10.1016/j.ins.2018.03.033
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The predictive performance of traditional supervised methods heavily depends on the amount of labeled data. However, obtaining labels is a difficult process in many real-life tasks, and only a small amount of labeled data is typically available for model learning. As an answer to this problem, the concept of semi-supervised learning has emerged. Semi supervised methods use unlabeled data in addition to labeled data to improve the performance of supervised methods. It is even more difficult to get labeled data for data mining problems with structured outputs since several labels need to be determined for each example. Multi-target regression (MTR) is one type of a structured output prediction problem, where we need to simultaneously predict multiple continuous variables. Despite the apparent need for semi supervised methods able to deal with MTR, only a few such methods are available and even those are difficult to use in practice and/or their advantages over supervised methods for MTR are not clear. This paper presents an extension of predictive clustering trees for MTR and ensembles thereof towards semi-supervised learning. The proposed method preserves the appealing characteristic of decision trees while enabling the use of unlabeled examples. In particular, the proposed semi-supervised trees for MTR are interpretable, easy to understand, fast to learn, and can handle both numeric and nominal descriptive features. We perform an extensive empirical evaluation in both an inductive and a transductive semi-supervised setting. The results show that the proposed method improves the performance of supervised predictive clustering trees and enhances their interpretability (due to reduced tree size), whereas, in the ensemble learning scenario, it outperforms its supervised counterpart in the transductive setting. The proposed methods have a mechanism for controlling the influence of unlabeled examples, which makes them highly useful in practice: This mechanism can protect them against a degradation of performance of their supervised counterparts-an inherent risk of semi-supervised learning. The proposed methods also outperform two existing semi-supervised methods for MTR. (C) 2018 Elsevier Inc. All rights reserved.
引用
收藏
页码:109 / 127
页数:19
相关论文
共 50 条
  • [41] Phenotype Prediction with Semi-supervised Classification Trees
    Levatic, Jurica
    Brbic, Maria
    Perdih, Tomaz Stepisnik
    Kocev, Dragi
    Vidulin, Vedrana
    Smuc, Tomislav
    Supek, Fran
    Dzeroski, Saso
    NEW FRONTIERS IN MINING COMPLEX PATTERNS, NFMCP 2017, 2018, 10785 : 138 - 150
  • [42] Semi-supervised target classification in multi-frequency echosounder data
    Choi, Changkyu
    Kampffmeyer, Michael
    Handegard, Nils Olav
    Salberg, Arnt-Borre
    Brautaset, Olav
    Eikvil, Line
    Jenssen, Robert
    ICES JOURNAL OF MARINE SCIENCE, 2021, 78 (07) : 2615 - 2627
  • [43] Ensemble learning with trees and rules: Supervised, semi-supervised, unsupervised
    Akdemir, Deniz
    Jannink, Jean-Luc
    INTELLIGENT DATA ANALYSIS, 2014, 18 (05) : 857 - 872
  • [44] Semi-Supervised Predictive Clustering Trees for (Hierarchical) Multi-Label Classification
    Levatic, Jurica
    Ceci, Michelangelo
    Kocev, Dragi
    Dzeroski, Saso
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2024, 2024
  • [45] SEMI-SUPERVISED HYPERSPECTRAL MANIFOLD LEARNING FOR REGRESSION
    Uto, Kuniaki
    Kosugi, Yukio
    Saito, Genya
    2015 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2015, : 9 - 12
  • [46] Semi-supervised contrastive regression for pharmaceutical processes
    Li, Yinlong
    Liao, Yilin
    Sun, Ziyue
    Liu, Xinggao
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [47] Learning Safe Prediction for Semi-Supervised Regression
    Li, Yu-Feng
    Zha, Han-Wen
    Zhou, Zhi-Hua
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2217 - 2223
  • [48] Huber Regression Analysis with a Semi-Supervised Method
    Wang, Yue
    Wang, Baobin
    Peng, Chaoquan
    Li, Xuefeng
    Yin, Hong
    MATHEMATICS, 2022, 10 (20)
  • [49] Analysis of Network Lasso for Semi-Supervised Regression
    Jung, A.
    Vesselinova, N.
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89 : 380 - 387
  • [50] Semi-supervised kernel methods for regression estimation
    Pozdnoukhov, Alexei
    Bengio, Samy
    2006 IEEE International Conference on Acoustics, Speech and Signal Processing, Vols 1-13, 2006, : 5435 - 5438