Task-Aware Dynamic Model Optimization for Multi-Task Learning

被引:0
|
作者
Choi, Sujin [1 ]
Jin, Hyundong [2 ]
Kim, Eunwoo [1 ,2 ]
机构
[1] Chung Ang Univ, Dept Artificial Intelligence, Seoul 06974, South Korea
[2] Chung Ang Univ, Sch Comp Sci & Engn, Seoul 06974, South Korea
关键词
Multi-task learning; resource-efficient learning; model optimization;
D O I
10.1109/ACCESS.2023.3339793
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-task learning (MTL) is a field in which a deep neural network simultaneously learns knowledge from multiple tasks. However, achieving resource-efficient MTL remains challenging due to entangled network parameters across tasks and varying task-specific complexity. Existing methods employ network compression techniques while maintaining comparable performance, but they often compress uniformly across all tasks without considering individual complexity. This can lead to suboptimal solutions due to entangled network parameters and memory inefficiency, as the parameters for each task may be insufficient or excessive. To address these challenges, we propose a framework called Dynamic Model Optimization (DMO) that dynamically allocates network parameters to groups based on task-specific complexity. This framework consists of three key steps: measuring task similarity and task difficulty, grouping tasks, and allocating parameters. This process involves the calculation of both weight and loss similarities across tasks and employs sample-wise loss as a measure of task difficulty. Tasks are grouped based on their similarities, and parameters are allocated with dynamic pruning according to task difficulty within their respective groups. We apply the proposed framework to MTL with various classification datasets. Experimental results demonstrate that the proposed approach achieves high performance while taking fewer network parameters than other MTL methods.
引用
收藏
页码:137709 / 137717
页数:9
相关论文
共 50 条
  • [21] Task-aware world model learning with meta weighting via bi-level optimization
    Yuan, Huining
    Dou, Hongkun
    Jiang, Xingyu
    Deng, Yue
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [22] A multi-task learning model with reinforcement optimization for ASD comorbidity discrimination
    Dong, Heyou
    Chen, Dan
    Chen, Yukang
    Tang, Yunbo
    Yin, Dingze
    Li, Xiaoli
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2024, 243
  • [23] Optimization of the Abstract Text Summarization Model Based on Multi-Task Learning
    Yao, Ben
    Ding, Gejian
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON ELECTRONIC INFORMATION TECHNOLOGY AND COMPUTER ENGINEERING, EITCE 2023, 2023, : 424 - 428
  • [24] Learning Sparse Task Relations in Multi-Task Learning
    Zhang, Yu
    Yang, Qiang
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2914 - 2920
  • [25] Task Variance Regularized Multi-Task Learning
    Mao, Yuren
    Wang, Zekai
    Liu, Weiwei
    Lin, Xuemin
    Hu, Wenbin
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (08) : 8615 - 8629
  • [26] Task Switching Network for Multi-task Learning
    Sun, Guolei
    Probst, Thomas
    Paudel, Danda Pani
    Popovic, Nikola
    Kanakis, Menelaos
    Patel, Jagruti
    Dai, Dengxin
    Van Gool, Luc
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8271 - 8280
  • [27] Learning to Teach Fairness-Aware Deep Multi-task Learning
    Roy, Arjun
    Ntoutsi, Eirini
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT I, 2023, 13713 : 710 - 726
  • [28] Task-Aware Query Recommendation
    Feild, Henry
    Allan, James
    SIGIR'13: THE PROCEEDINGS OF THE 36TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH & DEVELOPMENT IN INFORMATION RETRIEVAL, 2013, : 83 - 92
  • [29] Model-Protected Multi-Task Learning
    Liang, Jian
    Liu, Ziqi
    Zhou, Jiayu
    Jiang, Xiaoqian
    Zhang, Changshui
    Wang, Fei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (02) : 1002 - 1019
  • [30] Learning Task-Aware Energy Disaggregation: a Federated Approach
    Liu, Ruohong
    Chen, Yize
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 4412 - 4418