Task-Aware Dynamic Model Optimization for Multi-Task Learning

被引:0
|
作者
Choi, Sujin [1 ]
Jin, Hyundong [2 ]
Kim, Eunwoo [1 ,2 ]
机构
[1] Chung Ang Univ, Dept Artificial Intelligence, Seoul 06974, South Korea
[2] Chung Ang Univ, Sch Comp Sci & Engn, Seoul 06974, South Korea
关键词
Multi-task learning; resource-efficient learning; model optimization;
D O I
10.1109/ACCESS.2023.3339793
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-task learning (MTL) is a field in which a deep neural network simultaneously learns knowledge from multiple tasks. However, achieving resource-efficient MTL remains challenging due to entangled network parameters across tasks and varying task-specific complexity. Existing methods employ network compression techniques while maintaining comparable performance, but they often compress uniformly across all tasks without considering individual complexity. This can lead to suboptimal solutions due to entangled network parameters and memory inefficiency, as the parameters for each task may be insufficient or excessive. To address these challenges, we propose a framework called Dynamic Model Optimization (DMO) that dynamically allocates network parameters to groups based on task-specific complexity. This framework consists of three key steps: measuring task similarity and task difficulty, grouping tasks, and allocating parameters. This process involves the calculation of both weight and loss similarities across tasks and employs sample-wise loss as a measure of task difficulty. Tasks are grouped based on their similarities, and parameters are allocated with dynamic pruning according to task difficulty within their respective groups. We apply the proposed framework to MTL with various classification datasets. Experimental results demonstrate that the proposed approach achieves high performance while taking fewer network parameters than other MTL methods.
引用
收藏
页码:137709 / 137717
页数:9
相关论文
共 50 条
  • [1] AdaTask: A Task-Aware Adaptive Learning Rate Approach to Multi-Task Learning
    Yang, Enneng
    Pan, Junwei
    Wang, Ximei
    Yu, Haibin
    Shen, Li
    Chen, Xihua
    Xiao, Lei
    Jiang, Jie
    Guo, Guibing
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10745 - 10753
  • [2] Task-aware asynchronous multi-task model with class incremental contrastive learning for surgical scene understanding
    Lalithkumar Seenivasan
    Mobarakol Islam
    Mengya Xu
    Chwee Ming Lim
    Hongliang Ren
    [J]. International Journal of Computer Assisted Radiology and Surgery, 2023, 18 : 921 - 928
  • [3] Task-aware asynchronous multi-task model with class incremental contrastive learning for surgical scene understanding
    Seenivasan, Lalithkumar
    Islam, Mobarakol
    Xu, Mengya
    Lim, Chwee Ming
    Ren, Hongliang
    [J]. INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2023, 18 (05) : 921 - 928
  • [4] Task-Aware Optimization of Dynamic Fractional Permissions
    Angerer, Christoph M.
    [J]. LANGUAGES AND COMPILERS FOR PARALLEL COMPUTING, LCPC 2013, 2014, 8664 : 39 - 54
  • [5] Multi-task Graph Neural Architecture Search with Task-aware Collaboration and Curriculum
    Qin, Yijian
    Wang, Xin
    Zhang, Ziwei
    Chen, Hong
    Zhu, Wenwu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] TASK AWARE MULTI-TASK LEARNING FOR SPEECH TO TEXT TASKS
    Indurthi, Sathish
    Zaidi, Mohd Abbas
    Lakumarapu, Nikhil Kumar
    Lee, Beomseok
    Han, Hyojung
    Ahn, Seokchan
    Kim, Sangha
    Kim, Chanwoo
    Hwang, Inchul
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7723 - 7727
  • [7] Task-aware network: Mitigation of task-aware and task-free performance gap in online continual learning
    Hong, Yongwon
    Park, Sungho
    Byun, Hyeran
    [J]. NEUROCOMPUTING, 2023, 552
  • [8] Distractor-Aware Tracking with Multi-Task and Dynamic Feature Learning
    Liu, Weichun
    Tang, Xiaoan
    Zhao, Chenglin
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2021, 30 (02)
  • [9] Scale-Aware Task Message Transferring for Multi-Task Learning
    Sirejiding, Shalayiding
    Lu, Yuxiang
    Lu, Hongtao
    Ding, Yue
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1859 - 1864
  • [10] Learning Instance and Task-Aware Dynamic Kernels for Few-Shot Learning
    Ma, Rongkai
    Fang, Pengfei
    Avraham, Gil
    Zuo, Yan
    Zhu, Tianyu
    Drummond, Tom
    Harandi, Mehrtash
    [J]. COMPUTER VISION, ECCV 2022, PT XX, 2022, 13680 : 257 - 274