Multi-task learning for historical text normalization: Size matters

被引:0
|
作者
Bollmann, Marcel [1 ]
Sogaard, Anders [1 ]
Bingel, Joachim [1 ]
机构
[1] Univ Copenhagen, Dept Comp Sci, Copenhagen, Denmark
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Historical text normalization suffers from small datasets that exhibit high variance, and previous work has shown that multi-task learning can be used to leverage data from related problems in order to obtain more robust models. Previous work has been limited to datasets from a specific language and a specific historical period, and it is not clear whether results generalize. It therefore remains an open problem, when historical text normalization benefits from multi-task learning. We explore the benefits of multi-task learning across 10 different datasets, representing different languages and periods. Our main finding-contrary to what has been observed for other NLP tasks-is that multi-task learning mainly works when target task data is very scarce.
引用
收藏
页码:19 / 24
页数:6
相关论文
共 50 条
  • [1] Adversarial Multi-task Learning for Text Classification
    Liu, Pengfei
    Qiu, Xipeng
    Huang, Xuanjing
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1 - 10
  • [2] Generative Multi-Task Learning for Text Classification
    Zhao, Wei
    Gao, Hui
    Chen, Shuhui
    Wang, Nan
    [J]. IEEE ACCESS, 2020, 8 : 86380 - 86387
  • [3] TASK AWARE MULTI-TASK LEARNING FOR SPEECH TO TEXT TASKS
    Indurthi, Sathish
    Zaidi, Mohd Abbas
    Lakumarapu, Nikhil Kumar
    Lee, Beomseok
    Han, Hyojung
    Ahn, Seokchan
    Kim, Sangha
    Kim, Chanwoo
    Hwang, Inchul
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7723 - 7727
  • [4] Neural Multi-Task Text Normalization and Sanitization with Pointer-Generator
    Nguyen, Van-Hoang
    Sandro, Cavallari
    [J]. FIRST WORKSHOP ON NATURAL LANGUAGE INTERFACES, 2020, : 37 - 47
  • [5] ADVERSARIAL MULTI-TASK LEARNING FOR SPEAKER NORMALIZATION IN REPLAY DETECTION
    Suthokumar, Gajan
    Sethu, Vidhyasaharan
    Sriskandaraja, Kaavya
    Ambikairajah, Eliathamby
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6609 - 6613
  • [6] MCapsNet: Capsule Network for Text with Multi-Task Learning
    Xiao, Liqiang
    Zhang, Honglun
    Chen, Wenqing
    Wang, Yongkun
    Jin, Yaohui
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4565 - 4574
  • [7] A Multi-Task Learning Framework for Abstractive Text Summarization
    Lu, Yao
    Liu, Linqing
    Jiang, Zhile
    Yang, Min
    Goebel, Randy
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9987 - 9988
  • [8] Adaptive multi-task learning for speech to text translation
    Feng, Xin
    Zhao, Yue
    Zong, Wei
    Xu, Xiaona
    [J]. EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2024, 2024 (01):
  • [9] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    [J]. Memetic Computing, 2020, 12 : 355 - 369
  • [10] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    [J]. MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369