AUTOSEM: Automatic Task Selection and Mixing in Multi-Task Learning

被引:0
|
作者
Guo, Han [1 ]
Pasunuru, Ramakanth [1 ]
Bansal, Mohit [1 ]
机构
[1] Univ N Carolina, Chapel Hill, NC 27599 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-task learning (MTL) has achieved success over a wide range of problems, where the goal is to improve the performance of a primary task using a set of relevant auxiliary tasks. However, when the usefulness of the auxiliary tasks w.r.t. the primary task is not known a priori, the success of MTL models depends on the correct choice of these auxiliary tasks and also a balanced mixing ratio of these tasks during alternate training. These two problems could be resolved via manual intuition or hyper-parameter tuning over all combinatorial task choices, but this introduces inductive bias or is not scalable when the number of candidate auxiliary tasks is very large. To address these issues, we present AUTOSEM, a two-stage MTL pipeline, where the first stage automatically selects the most useful auxiliary tasks via a Beta-Bernoulli multi-armed bandit with Thompson Sampling, and the second stage learns the training mixing ratio of these selected auxiliary tasks via a Gaussian Process based Bayesian optimization framework. We conduct several MTL experiments on the GLUE language understanding tasks, and show that our AUTOSEM framework can successfully find relevant auxiliary tasks and automatically learn their mixing ratio, achieving significant performance boosts on several primary tasks. Finally, we present ablations for each stage of AUTOSEM and analyze the learned auxiliary task choices.
引用
收藏
页码:3520 / 3531
页数:12
相关论文
共 50 条
  • [1] Variable Selection and Task Grouping for Multi-Task Learning
    Jeong, Jun-Yong
    Jun, Chi-Hyuck
    [J]. KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, : 1589 - 1598
  • [2] Multi-task learning for stock selection
    Ghosn, J
    Bengio, Y
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 9: PROCEEDINGS OF THE 1996 CONFERENCE, 1997, 9 : 946 - 952
  • [3] Automatic Temporal Relation in Multi-Task Learning
    Zhou, Menghui
    Yang, Po
    [J]. PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 3570 - 3580
  • [4] Automatic Cataract Detection with Multi-Task Learning
    Wu, Hongjie
    Lv, Jiancheng
    Wang, Jian
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [5] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    [J]. Memetic Computing, 2020, 12 : 355 - 369
  • [6] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    [J]. MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [7] Structured feature selection and task relationship inference for multi-task learning
    Hongliang Fei
    Jun Huan
    [J]. Knowledge and Information Systems, 2013, 35 : 345 - 364
  • [8] Structured feature selection and task relationship inference for multi-task learning
    Fei, Hongliang
    Huan, Jun
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2013, 35 (02) : 345 - 364
  • [9] Automatic Structural Search for Multi-task Learning VALPs
    Garciarena, Unai
    Mendiburu, Alexander
    Santana, Roberto
    [J]. OPTIMIZATION AND LEARNING, 2020, 1173 : 25 - 36
  • [10] Automatic Expert Selection for Multi-Scenario and Multi-Task Search
    Zou, Xinyu
    Hu, Zhi
    Zhao, Yiming
    Ding, Xuchu
    Liu, Zhongyi
    Li, Chenliang
    Sun, Aixin
    [J]. PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1535 - 1544