Multi-task Active Learning for Pre-trained Transformer-based Models

被引:6
|
作者
Rotman, Guy [1 ]
Reichart, Roi [1 ]
机构
[1] Fac Ind Engn & Management, Technion, IIT, Haifa, Israel
关键词
Computational linguistics - Iterative methods - Learning algorithms - Learning systems;
D O I
10.1162/tacl_a_00515
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-task learning, in which several tasks are jointly learned by a single model, allows NLP models to share information from multiple annotations and may facilitate better predictions when the tasks are inter-related. This technique, however, requires annotating the same text with multiple annotation schemes, which may be costly and laborious. Active learning (AL) has been demonstrated to optimize annotation processes by iteratively selecting unlabeled examples whose annotation is most valuable for the NLP model. Yet, multi-task active learning (MT-AL) has not been applied to state-of-the-art pre-trained Transformer-based NLP models. This paper aims to close this gap. We explore various multi-task selection criteria in three realistic multi-task scenarios, reflecting different relations between the participating tasks, and demonstrate the effectiveness of multi-task compared to single-task selection. Our results suggest that MT-AL can be effectively used in order to minimize annotation efforts for multi-task NLP models.(1)
引用
收藏
页码:1209 / 1228
页数:20
相关论文
共 50 条
  • [31] MASTER: Multi-task Pre-trained Bottlenecked Masked Autoencoders Are Better Dense Retrievers
    Zhou, Kun
    Liu, Xiao
    Gong, Yeyun
    Zhao, Wayne Xin
    Jiang, Daxin
    Duan, Nan
    Wen, Ji-Rong
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 : 630 - 647
  • [32] PTMB: An online satellite task scheduling framework based on pre-trained Markov decision process for multi-task scenario
    Li, Guohao
    Li, Xuefei
    Li, Jing
    Chen, Jia
    Shen, Xin
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 284
  • [33] PARFormer: Transformer-Based Multi-Task Network for Pedestrian Attribute Recognition
    Fan, Xinwen
    Zhang, Yukang
    Lu, Yang
    Wang, Hanzi
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) : 411 - 423
  • [34] Automatic Title Generation for Learning Resources and Pathways with Pre-trained Transformer Models
    Mishra, Prakhar
    Diwan, Chaitali
    Srinivasa, Srinath
    Srinivasaraghavan, G.
    [J]. INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2021, 15 (04) : 487 - 510
  • [35] On the effect of dropping layers of pre-trained transformer models
    Sajjad, Hassan
    Dalvi, Fahim
    Durrani, Nadir
    Nakov, Preslav
    [J]. COMPUTER SPEECH AND LANGUAGE, 2022, 77
  • [36] Automatic Question Generation using RNN-based and Pre-trained Transformer-based Models in Low Resource Indonesian Language
    Vincentio, Karissa
    Suhartono, Derwin
    [J]. INFORMATICA-AN INTERNATIONAL JOURNAL OF COMPUTING AND INFORMATICS, 2022, 46 (07): : 103 - 118
  • [37] Framing and BERTology: A Data-Centric Approach to Integration of Linguistic Features into Transformer-Based Pre-trained Language Models
    Avetisyan, Hayastan
    Safikhani, Parisa
    Broneske, David
    [J]. INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 4, INTELLISYS 2023, 2024, 825 : 81 - 90
  • [38] Multi-task learning models for predicting active compounds
    Zhao, Zhili
    Qin, Jian
    Gou, Zhuoyue
    Zhang, Yanan
    Yang, Yi
    [J]. JOURNAL OF BIOMEDICAL INFORMATICS, 2020, 108
  • [39] A Transformer-Based Multi-Task Learning Framework for Myoelectric Pattern Recognition Supporting Muscle Force Estimation
    Li, Xinhui
    Zhang, Xu
    Zhang, Liwei
    Chen, Xiang
    Zhou, Ping
    [J]. IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2023, 31 : 3255 - 3264
  • [40] Photo-based Carbohydrates Counting using Pre-trained Transformer Models
    Contreras, Ivan
    Guso, Marti
    Beneyto, Aleix
    Vehi, Josep
    [J]. IFAC PAPERSONLINE, 2023, 56 (02): : 11533 - 11538