Adaptive Hard Parameter Sharing Method Based on Multi-Task Deep Learning

被引:1
|
作者
Wang, Hongxia [1 ]
Jin, Xiao [1 ]
Du, Yukun [1 ]
Zhang, Nan [1 ]
Hao, Hongxia [1 ]
机构
[1] Nanjing Audit Univ, Sch Stat & Data Sci, Nanjing 211815, Peoples R China
关键词
multi-task learning; continuous gradient difference threshold; warm-up; training iteration threshold; information sharing; adaptive nodes; NEURAL-NETWORK; MODEL;
D O I
10.3390/math11224639
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Multi-task learning (MTL) improves the performance achieved on each task by exploiting the relevant information between tasks. At present, most of the mainstream deep MTL models are based on hard parameter sharing mechanisms, which can reduce the risk of model overfitting. However, negative knowledge transfer may occur, which hinders the performance improvement achieved for each task. In this paper, for situations when multiple tasks are jointly trained, we propose the adaptive hard parameter sharing method. On the basis of the adaptive hard parameter sharing method, the number of nodes in the network is dynamically updated by setting a continuous gradient difference-based sign threshold and a warm-up training iteration threshold through the relationships between the parameters and the loss function. After each task fully utilizes the shared information, adaptive nodes are used to further optimize each task, reducing the impact of negative migration. By using simulation studies and instance analyses, we demonstrate theoretical proof that the performance of the proposed method is better than that of the competing method.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Task Adaptive Parameter Sharing for Multi-Task Learning
    Wallingford, Matthew
    Li, Hao
    Achille, Alessandro
    Ravichandran, Avinash
    Fowlkes, Charless
    Bhotika, Rahul
    Soatto, Stefano
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7551 - 7560
  • [2] Multi-Task Learning for Sentiment Analysis with Hard-Sharing and Task Recognition Mechanisms
    Zhang, Jian
    Yan, Ke
    Mo, Yuchang
    [J]. INFORMATION, 2021, 12 (05)
  • [3] A multi-task deep learning based vulnerability severity prediction method
    Shan, Chun
    Zhang, Ziyi
    Zhou, Siyi
    [J]. 2023 IEEE 12TH INTERNATIONAL CONFERENCE ON CLOUD NETWORKING, CLOUDNET, 2023, : 307 - 315
  • [4] Fitting and sharing multi-task learning
    Piao, Chengkai
    Wei, Jinmao
    [J]. APPLIED INTELLIGENCE, 2024, 54 (9-10) : 6918 - 6929
  • [5] An adaptive multi-objective multi-task scheduling method by hierarchical deep reinforcement learning
    Zhang, Jianxiong
    Guo, Bing
    Ding, Xuefeng
    Hu, Dasha
    Tang, Jun
    Du, Ke
    Tang, Chao
    Jiang, Yuming
    [J]. APPLIED SOFT COMPUTING, 2024, 154
  • [6] Multi-Task Learning Using BERT With Soft Parameter Sharing Between Layers
    Pahari, Niraj
    Shimada, Kazutaka
    [J]. 2022 JOINT 12TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS AND 23RD INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS (SCIS&ISIS), 2022,
  • [7] Robust Estimator based Adaptive Multi-Task Learning
    Zhu, Peiyuan
    Chen, Cailian
    He, Jianping
    Zhu, Shanying
    [J]. 2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 740 - 747
  • [8] Gearbox fault diagnosis method based on deep learning multi-task framework
    Chen, Yao
    Liang, Ruijun
    Ran, Wenfeng
    Chen, Weifang
    [J]. INTERNATIONAL JOURNAL OF STRUCTURAL INTEGRITY, 2023, 14 (03) : 401 - 415
  • [9] Multi-Task Water Quality Colorimetric Detection Method Based on Deep Learning
    Zhang, Shenlan
    Wu, Shaojie
    Chen, Liqiang
    Guo, Pengxin
    Jiang, Xincheng
    Pan, Hongcheng
    Li, Yuhong
    [J]. Sensors, 2024, 24 (22)
  • [10] Multi-Adaptive Optimization for multi-task learning with deep neural networks
    Hervella, alvaro S.
    Rouco, Jose
    Novo, Jorge
    Ortega, Marcos
    [J]. NEURAL NETWORKS, 2024, 170 : 254 - 265