On Partial Multi-Task Learning

被引:4
|
作者
He, Yi [1 ]
Wu, Baijun [1 ]
Wu, Di [2 ]
Wu, Xindong [3 ,4 ]
机构
[1] Univ Louisiana Lafayette, Sch Comp & Informat, Lafayette, LA 70504 USA
[2] Chinese Acad Sci, Chongqing Inst Green & Intelligent Technol, Beijing, Peoples R China
[3] Mininglamp Acad Sci, Mininglamp Technol, Beijing, Peoples R China
[4] Hefei Univ Technol, Minist Educ, Key Lab Knowledge Engn Big Data, Hefei, Peoples R China
基金
美国国家科学基金会;
关键词
MATRIX COMPLETION; CLASSIFICATION;
D O I
10.3233/FAIA200216
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-Task Learning (MTL) has shown its effectiveness in real applications where many related tasks could be handled together. Existing MTL methods make predictions for multiple tasks based on the data examples of the corresponding tasks. However, the data examples of some tasks are expensive or time-consuming to collect in practice, which reduces the applicability of MTL. For example, a patient may be asked to provide her microtome test reports and MRI images for illness diagnosis in MTL-based system [37,40]. It would be valuable if MTL can predict the abnormalities for such medical tests by feeding with some easy-to-collect data examples from other related tests instead of directly collecting data examples from them. We term such a new paradigm as multi-task learning from partial examples. The challenges of partial multi-task learning are twofold. First, the data examples from different tasks may be represented in different feature spaces. Second, the data examples could be incomplete for predicting the labels of all tasks. To overcome these challenges, we in this paper propose a novel algorithm, named Generative Learning with Partial Multi-Tasks (GPMT). The key idea of GPMT is to discover a shared latent feature space that harmonizes disparate feature information of multiple tasks. Given a partial example, the information contained in its missing feature representations is recovered by projecting it onto the latent space. A learner trained on the latent space then enjoys complete information included in the original features and the recovered missing features, and thus can predict the labels for the partial examples. Our theoretical analysis shows that the GPMT guarantees a performance gain comparing with training an individual learner for each task. Extensive experiments demonstrate the superiority of GPMT on both synthetic and real datasets.
引用
收藏
页码:1174 / 1181
页数:8
相关论文
共 50 条
  • [41] Polymer informatics with multi-task learning
    Kuenneth, Christopher
    Rajan, Arunkumar Chitteth
    Tran, Huan
    Chen, Lihua
    Kim, Chiho
    Ramprasad, Rampi
    [J]. PATTERNS, 2021, 2 (04):
  • [42] Gradient Surgery for Multi-Task Learning
    Yu, Tianhe
    Kumar, Saurabh
    Gupta, Abhishek
    Levine, Sergey
    Hausman, Karol
    Finn, Chelsea
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [43] Multi-task Learning for Recommender Systems
    Ning, Xia
    Karypis, George
    [J]. PROCEEDINGS OF 2ND ASIAN CONFERENCE ON MACHINE LEARNING (ACML2010), 2010, 13 : 269 - 284
  • [44] Stock Ranking with Multi-Task Learning
    Ma, Tao
    Tan, Ying
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2022, 199
  • [45] ADAPTIVE AND ROBUST MULTI-TASK LEARNING
    Duan, Yaqi
    Wang, Kaizheng
    [J]. ANNALS OF STATISTICS, 2023, 51 (05): : 2015 - 2039
  • [46] Bounds for linear multi-task learning
    Maurer, A
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2006, 7 : 117 - 139
  • [47] Multi-task learning for pKa prediction
    Skolidis, Grigorios
    Hansen, Katja
    Sanguinetti, Guido
    Rupp, Matthias
    [J]. JOURNAL OF COMPUTER-AIDED MOLECULAR DESIGN, 2012, 26 (07) : 883 - 895
  • [48] Federated Multi-task Graph Learning
    Liu, Yijing
    Han, Dongming
    Zhang, Jianwei
    Zhu, Haiyang
    Xu, Mingliang
    Chen, Wei
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2022, 13 (05)
  • [49] Multi-task and Lifelong Learning of Kernels
    Pentina, Anastasia
    Ben-David, Shai
    [J]. ALGORITHMIC LEARNING THEORY, ALT 2015, 2015, 9355 : 194 - 208
  • [50] Pareto Multi-task Deep Learning
    Riccio, Salvatore D.
    Dyankov, Deyan
    Jansen, Giorgio
    Di Fatta, Giuseppe
    Nicosia, Giuseppe
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 132 - 141