Efficient Computation Sharing for Multi-Task Visual Scene Understanding

被引:0
|
作者
Shoouri, Sara [1 ]
Yang, Mingyu [1 ]
Fan, Zichen [1 ]
Kim, Hun-Seok [1 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
关键词
MODEL;
D O I
10.1109/ICCV51070.2023.01571
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Solving multiple visual tasks using individual models can be resource-intensive, while multi-task learning can conserve resources by sharing knowledge across different tasks. Despite the benefits of multi-task learning, such techniques can struggle with balancing the loss for each task, leading to potential performance degradation. We present a novel computation- and parameter-sharing framework that balances efficiency and accuracy to perform multiple visual tasks utilizing individually-trained single-task transformers. Our method is motivated by transfer learning schemes to reduce computational and parameter storage costs while maintaining the desired performance. Our approach involves splitting the tasks into a base task and the other sub-tasks, and sharing a significant portion of activations and parameters/weights between the base and sub-tasks to decrease inter-task redundancies and enhance knowledge sharing. The evaluation conducted on NYUD-v2 and PASCAL-context datasets shows that our method is superior to the state-of-the-art transformer-based multi-task learning techniques with higher accuracy and reduced computational resources. Moreover, our method is extended to video stream inputs, further reducing computational costs by efficiently sharing information across the temporal domain as well as the task domain. Our codes are available at https://github.com/sarashoouri/EfficientMTL.
引用
收藏
页码:17084 / 17095
页数:12
相关论文
共 50 条
  • [1] InvPT plus plus : Inverted Pyramid Multi-Task Transformer for Visual Scene Understanding
    Ye, Hanrong
    Xu, Dan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 7493 - 7508
  • [2] Multi-Task Deep Learning Design and Training Tool for Unified Visual Driving Scene Understanding
    Won, Woong-Jae
    Kim, Tae Hun
    Kwon, Soon
    2019 19TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2019), 2019, : 356 - 360
  • [3] HirMTL: Hierarchical Multi-Task Learning for dense scene understanding
    Luo, Huilan
    Hu, Weixia
    Wei, Yixiao
    He, Jianlong
    Yu, Minghao
    NEURAL NETWORKS, 2025, 181
  • [4] Inverted Pyramid Multi-task Transformer for Dense Scene Understanding
    Ye, Hanrong
    Xu, Dan
    COMPUTER VISION - ECCV 2022, PT XXVII, 2022, 13687 : 514 - 530
  • [5] VMT-Adapter: Parameter-Efficient Transfer Learning for Multi-Task Dense Scene Understanding
    Xin, Yi
    Du, Junlong
    Wang, Qiang
    Lin, Zhiwen
    Yan, Ke
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 16085 - 16093
  • [6] Efficient Multi-Task Scene Analysis with RGB-D Transformers
    Fischedick, Soehnke Benedikt
    Seichter, Daniel
    Schmidt, Robin
    Rabes, Leonard
    Gross, Horst-Michael
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [7] MT-UNET: A NOVEL U-NET BASED MULTI-TASK ARCHITECTURE FOR VISUAL SCENE UNDERSTANDING
    Jha, Ankit
    Kumar, Awanish
    Pande, Shivam
    Banerjee, Biplab
    Chaudhuri, Subhasis
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2191 - 2195
  • [8] Research on Road Scene Understanding of Autonomous Vehicles Based on Multi-Task Learning
    Guo, Jinghua
    Wang, Jingyao
    Wang, Huinian
    Xiao, Baoping
    He, Zhifei
    Li, Lubin
    SENSORS, 2023, 23 (13)
  • [9] Global-Reasoned Multi-Task Learning Model for Surgical Scene Understanding
    Seenivasan, Lalithkumar
    Mitheran, Sai
    Islam, Mobarakol
    Ren, Hongliang
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3858 - 3865
  • [10] Visual Person Understanding Through Multi-task and Multi-dataset Learning
    Pfeiffer, Kilian
    Hermans, Alexander
    Sarandi, Istvan
    Weber, Mark
    Leibe, Bastian
    PATTERN RECOGNITION, DAGM GCPR 2019, 2019, 11824 : 551 - 566