Boosting for regression transfer via importance sampling

被引:3
|
作者
Gupta, Shrey [1 ]
Bi, Jianzhao [2 ]
Liu, Yang [3 ]
Wildani, Avani [1 ]
机构
[1] Emory Univ, Dept Comp Sci, Atlanta, GA 30322 USA
[2] Univ Washington, Dept Environm & Occupat Hlth Sci, Seattle, WA USA
[3] Emory Univ, Gangarosa Dept Environm Hlth, Atlanta, GA USA
基金
美国国家卫生研究院;
关键词
Instance transfer learning; Negative transfer; Domain adaptation; COMPLEXITY; FEATURES;
D O I
10.1007/s41060-023-00414-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Current instance transfer learning (ITL) methodologies use domain adaptation and sub-space transformation to achieve successful transfer learning. However, these methodologies, in their processes, sometimes overfit on the target dataset or suffer from negative transfer if the test dataset has a high variance. Boosting methodologies have been shown to reduce the risk of overfitting by iteratively re-weighing instances with high-residual. However, this balance is usually achieved with parameter optimization, as well as reducing the skewness in weights produced due to the size of the source dataset. While the former can be achieved, the latter is more challenging and can lead to negative transfer. We introduce a simpler and more robust fix to this problem by building upon the popular boosting ITL regression methodology, two-stage TrAdaBoost.R2. Our methodology, S-TrAdaBoost.R2, is a boosting-based ensemble methodology that utilizes importance sampling to reduce the skewness due to the source dataset. We show that S-TrAdaBoost.R2 performs better than competitive transfer learning methodologies 63% of the time. It also displays consistency in its performance over diverse datasets with varying complexities, as opposed to the sporadic results observed for other transfer learning methodologies.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Transfer of Samples in Policy Search via Multiple Importance Sampling
    Tirinzoni, Andrea
    Salvini, Mattia
    Restelli, Marcello
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [2] CounterFactual Regression with Importance Sampling Weights
    Hassanpour, Negar
    Greiner, Russell
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5880 - 5887
  • [3] SCALAR ON NETWORK REGRESSION VIA BOOSTING
    Morris, Emily L.
    He, Kevin
    Kang, Jian
    ANNALS OF APPLIED STATISTICS, 2022, 16 (04): : 2755 - 2773
  • [4] IMPORTANCE SAMPLING VIA A SIMULACRUM
    WESSEL, AE
    HALL, EB
    WISE, GL
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 1990, 327 (05): : 771 - 783
  • [5] Robust boosting via self-sampling
    Liu, Xiaoshuang
    Luo, Senlin
    Pan, Limin
    Pan, Limin (panlimin2016@gmail.com), 1600, Elsevier B.V., Netherlands (193):
  • [6] Robust boosting via self-sampling
    Liu, Xiaoshuang
    Luo, Senlin
    Pan, Limin
    KNOWLEDGE-BASED SYSTEMS, 2020, 193
  • [7] NAUTILUS: boosting Bayesian importance nested sampling with deep learning
    Lange, Johannes U.
    MONTHLY NOTICES OF THE ROYAL ASTRONOMICAL SOCIETY, 2023, 525 (02) : 3181 - 3194
  • [8] Importance Sampling via Local Sensitivity
    Raj, Anant
    Musco, Cameron
    Mackey, Lester
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 3099 - 3108
  • [9] Importance sampling via the estimated sampler
    Henmi, Masayuki
    Yoshida, Ryo
    Eguchi, Shinto
    BIOMETRIKA, 2007, 94 (04) : 985 - 991
  • [10] Information Fusion via Importance Sampling
    Saucan, Augustin A.
    Elvira, Victor
    Varshney, Pramod K.
    Win, Moe Z.
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2024, 10 : 376 - 389