MTFormer: Multi-task Learning via Transformer and Cross-Task Reasoning

被引:8
|
作者
Xu, Xiaogang [1 ]
Zhao, Hengshuang [2 ,3 ]
Vineet, Vibhav [4 ]
Lim, Ser-Nam [5 ]
Torralba, Antonio [2 ]
机构
[1] CUHK, Hong Kong, Peoples R China
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[3] HKU, Hong Kong, Peoples R China
[4] Microsoft Res, Redmond, WA USA
[5] Meta AI, New York, NY USA
来源
关键词
Multi-task learning; Transformer; Cross-task reasoning;
D O I
10.1007/978-3-031-19812-0_18
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we explore the advantages of utilizing transformer structures for addressing multi-task learning (MTL). Specifically, we demonstrate that models with transformer structures are more appropriate for MTL than convolutional neural networks (CNNs), and we propose a novel transformer-based architecture named MTFormer for MTL. In the framework, multiple tasks share the same transformer encoder and transformer decoder, and lightweight branches are introduced to harvest task-specific outputs, which increases the MTL performance and reduces the time-space complexity. Furthermore, information from different task domains can benefit each other, and we conduct cross-task reasoning. We propose a cross-task attention mechanism for further boosting the MTL results. The cross-task attention mechanism brings little parameters and computations while introducing extra performance improvements. Besides, we design a self-supervised cross-task contrastive learning algorithm for further boosting the MTL performance. Extensive experiments are conducted on two multi-task learning datasets, on which MTFormer achieves state-of-the-art results with limited network parameters and computations. It also demonstrates significant superiorities for few-shot learning and zero-shot learning.
引用
收藏
页码:304 / 321
页数:18
相关论文
共 50 条
  • [41] Object localization via evaluation multi-task learning
    Tian, Yan
    Wang, Huiyan
    Wang, Xun
    NEUROCOMPUTING, 2017, 253 : 34 - 41
  • [42] Multi-task learning via linear functional strategy
    Rastogi, Abhishake
    Sampath, Sivananthan
    JOURNAL OF COMPLEXITY, 2017, 43 : 51 - 75
  • [43] Knowledge triple mining via multi-task learning
    Zhang, Zhao
    Zhuang, Fuzhen
    Li, Xuebing
    Niu, Zheng-Yu
    He, Jia
    He, Qing
    Xiong, Hui
    INFORMATION SYSTEMS, 2019, 80 : 64 - 75
  • [44] Multi-task learning regression via convex clustering
    Okazaki, Akira
    Kawano, Shuichi
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2024, 195
  • [45] Binaural Audio Generation via Multi-task Learning
    Li, Sijia
    Liu, Shiguang
    Manocha, Dinesh
    ACM TRANSACTIONS ON GRAPHICS, 2021, 40 (06):
  • [46] Boosted multi-task learning
    Olivier Chapelle
    Pannagadatta Shivaswamy
    Srinivas Vadrevu
    Kilian Weinberger
    Ya Zhang
    Belle Tseng
    Machine Learning, 2011, 85 : 149 - 173
  • [47] An overview of multi-task learning
    Zhang, Yu
    Yang, Qiang
    NATIONAL SCIENCE REVIEW, 2018, 5 (01) : 30 - 43
  • [48] On-Device Deep Multi-Task Inference via Multi-Task Zipping
    He, Xiaoxi
    Wang, Xu
    Zhou, Zimu
    Wu, Jiahang
    Yang, Zheng
    Thiele, Lothar
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (05) : 2878 - 2891
  • [49] On Partial Multi-Task Learning
    He, Yi
    Wu, Baijun
    Wu, Di
    Wu, Xindong
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1174 - 1181
  • [50] Calibrated Multi-Task Learning
    Nie, Feiping
    Hu, Zhanxuan
    Li, Xuelong
    KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, : 2012 - 2021