A new transfer learning framework with application to model-agnostic multi-task learning

被引:7
|
作者
Gupta, Sunil [1 ]
Rana, Santu [1 ]
Saha, Budhaditya [1 ]
Phung, Dinh [1 ]
Venkatesh, Svetha [1 ]
机构
[1] Deakin Univ, Ctr Pattern Recognit & Data Analyt PRaDA, Geelong Waurn Ponds Campus, Waurn Ponds, Vic, Australia
关键词
Multi-task learning; Model-agnostic framework; Meta algorithm; Classification; Regression; CLASSIFICATION;
D O I
10.1007/s10115-016-0926-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from small number of examples is a challenging problem in machine learning. An effective way to improve the performance is through exploiting knowledge from other related tasks. Multi-task learning (MTL) is one such useful paradigm that aims to improve the performance through jointly modeling multiple related tasks. Although there exist numerous classification or regression models in machine learning literature, most of the MTL models are built around ridge or logistic regression. There exist some limited works, which propose multi-task extension of techniques such as support vector machine, Gaussian processes. However, all these MTL models are tied to specific classification or regression algorithms and there is no single MTL algorithm that can be used at a meta level for any given learning algorithm. Addressing this problem, we propose a generic, model-agnostic joint modeling framework that can take any classification or regression algorithm of a practitioner's choice (standard or custom-built) and build its MTL variant. The key observation that drives our framework is that due to small number of examples, the estimates of task parameters are usually poor, and we show that this leads to an under-estimation of task relatedness between any two tasks with high probability. We derive an algorithm that brings the tasks closer to their true relatedness by improving the estimates of task parameters. This is achieved by appropriate sharing of data across tasks. We provide the detail theoretical underpinning of the algorithm. Through our experiments with both synthetic and real datasets, we demonstrate that the multi-task variants of several classifiers/regressors (logistic regression, support vector machine, K-nearest neighbor, Random Forest, ridge regression, support vector regression) convincingly outperform their single-task counterparts. We also show that the proposed model performs comparable or better than many state-of-the-art MTL and transfer learning baselines.
引用
收藏
页码:933 / 973
页数:41
相关论文
共 50 条
  • [31] Focused multi-task learning in a Gaussian process framework
    Gayle Leen
    Jaakko Peltonen
    Samuel Kaski
    Machine Learning, 2012, 89 : 157 - 182
  • [32] A Multi-Task Learning Framework for Abstractive Text Summarization
    Lu, Yao
    Liu, Linqing
    Jiang, Zhile
    Yang, Min
    Goebel, Randy
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9987 - 9988
  • [33] Learning to Branch for Multi-Task Learning
    Guo, Pengsheng
    Lee, Chen-Yu
    Ulbricht, Daniel
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [34] Focused multi-task learning in a Gaussian process framework
    Leen, Gayle
    Peltonen, Jaakko
    Kaski, Samuel
    MACHINE LEARNING, 2012, 89 (1-2) : 157 - 182
  • [35] A multi-task framework for metric learning with common subspace
    Peipei Yang
    Kaizhu Huang
    Cheng-Lin Liu
    Neural Computing and Applications, 2013, 22 : 1337 - 1347
  • [36] A Multi-task Learning Framework for Opinion Triplet Extraction
    Zhang, Chen
    Li, Qiuchi
    Song, Dawei
    Wang, Benyou
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 819 - 828
  • [37] Online Multi-Task Learning Framework for Ensemble Forecasting
    Xu, Jianpeng
    Tan, Pang-Ning
    Zhou, Jiayu
    Luo, Lifeng
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2017, 29 (06) : 1268 - 1280
  • [38] A Multi-task Learning Framework for Product Ranking with BERT
    Wu, Xuyang
    Magnani, Alessandro
    Chaidaroon, Suthee
    Puthenputhussery, Ajit
    Liao, Ciya
    Fang, Yi
    PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 493 - 501
  • [39] Learning to Branch for Multi-Task Learning
    Guo, Pengsheng
    Lee, Chen-Yu
    Ulbricht, Daniel
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [40] Revisiting Model-Agnostic Private Learning: Faster Rates and Active Learning
    Liu, Chong
    Zhu, Yuqing
    Chaudhuri, Kamalika
    Wang, Yu-Xiang
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130