Attentive Task Interaction Network for Multi-Task Learning

被引:2
|
作者
Sinodinos, Dimitrios [1 ]
Armanfard, Narges [1 ]
机构
[1] McGill Univ, Mila Quebec AI Inst, Dept Elect & Comp Engn, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
10.1109/ICPR56361.2022.9956614
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multitask learning (MTL) has recently gained a lot of popularity as a learning paradigm that can lead to improved per-task performance while also using fewer per-task model parameters compared to single task learning. One of the biggest challenges regarding MTL networks involves how to share features across tasks. To address this challenge, we propose the Attentive Task Interaction Network (ATI-Net). ATI-Net employs knowledge distillation of the latent features for each task, then combines the feature maps to provide improved contextualized information to the decoder. This novel approach to introducing knowledge distillation into an attention based multitask network outperforms state of the art MTL baselines such as the standalone MTAN and PAD-Net, with roughly the same number of model parameters.
引用
收藏
页码:2885 / 2891
页数:7
相关论文
共 50 条
  • [1] Attentive Multi-task Deep Reinforcement Learning
    Bram, Timo
    Brunner, Gino
    Richter, Oliver
    Wattenhofer, Roger
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT III, 2020, 11908 : 134 - 149
  • [2] Task Switching Network for Multi-task Learning
    Sun, Guolei
    Probst, Thomas
    Paudel, Danda Pani
    Popovic, Nikola
    Kanakis, Menelaos
    Patel, Jagruti
    Dai, Dengxin
    Van Gool, Luc
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8271 - 8280
  • [3] Attentive multi-task learning for group itinerary recommendation
    Lei Chen
    Jie Cao
    Huanhuan Chen
    Weichao Liang
    Haicheng Tao
    Guixiang Zhu
    [J]. Knowledge and Information Systems, 2021, 63 : 1687 - 1716
  • [4] Attentive multi-task learning for group itinerary recommendation
    Chen, Lei
    Cao, Jie
    Chen, Huanhuan
    Liang, Weichao
    Tao, Haicheng
    Zhu, Guixiang
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2021, 63 (07) : 1687 - 1716
  • [5] Multi-Task Feature Interaction Learning
    Lin, Kaixiang
    Xu, Jianpeng
    Baytas, Inci M.
    Ji, Shuiwang
    Zhou, Jiayu
    [J]. KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 1735 - 1744
  • [6] Multi-Task Network Representation Learning
    Xie, Yu
    Jin, Peixuan
    Gong, Maoguo
    Zhang, Chen
    Yu, Bin
    [J]. FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [7] Multi-Task Decouple Learning With Hierarchical Attentive Point Process
    Wu, Weichang
    Zhang, Xiaolu
    Zhao, Shiwan
    Fu, Chilin
    Zhou, Jun
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (04) : 1741 - 1757
  • [8] Co-Attentive Multi-Task Learning for Explainable Recommendation
    Chen, Zhongxia
    Wang, Xiting
    Xie, Xing
    Wu, Tong
    Bu, Guoqing
    Wang, Yining
    Chen, Enhong
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2137 - 2143
  • [9] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    [J]. Memetic Computing, 2020, 12 : 355 - 369
  • [10] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    [J]. MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369