Multi-Task Learning for Abstractive and Extractive Summarization

被引:29
|
作者
Chen, Yangbin [1 ]
Ma, Yun [1 ]
Mao, Xudong [2 ]
Li, Qing [2 ]
机构
[1] City Univ Hong Kong, Kowloon Tong, Hong Kong, Peoples R China
[2] Hong Kong Polytech Univ, Hung Hom, Hong Kong, Peoples R China
关键词
Automatic document summarization; Multi-task learning; Attention mechanism;
D O I
10.1007/s41019-019-0087-7
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The abstractive method and extractive method are two main approaches for automatic document summarization. In this paper, to fully integrate the relatedness and advantages of both approaches, we propose a general unified framework for abstractive summarization which incorporates extractive summarization as an auxiliary task. In particular, our framework is composed of a shared hierarchical document encoder, a hierarchical attention mechanism-based decoder, and an extractor. We adopt multi-task learning method to train these two tasks jointly, which enables the shared encoder to better capture the semantics of the document. Moreover, as our main task is abstractive summarization, we constrain the attention learned in the abstractive task with the labels of the extractive task to strengthen the consistency between the two tasks. Experiments on the CNN/DailyMail dataset demonstrate that both the auxiliary task and the attention constraint contribute to improve the performance significantly, and our model is comparable to the state-of-the-art abstractive models. In addition, we cut half number of labels of the extractive task, pretrain the extractor, and jointly train the two tasks using the estimated sentence salience of the extractive task to constrain the attention of the abstractive task. The results do not decrease much compared with using full-labeled data of the auxiliary task.
引用
收藏
页码:14 / 23
页数:10
相关论文
共 50 条
  • [1] Multi-Task Learning for Abstractive and Extractive Summarization
    Yangbin Chen
    Yun Ma
    Xudong Mao
    Qing Li
    [J]. Data Science and Engineering, 2019, 4 (1) : 14 - 23
  • [2] A Multi-Task Learning Framework for Abstractive Text Summarization
    Lu, Yao
    Liu, Linqing
    Jiang, Zhile
    Yang, Min
    Goebel, Randy
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9987 - 9988
  • [3] Multi-Task Learning for Cross-Lingual Abstractive Summarization
    Takase, Sho
    Okazaki, Naoaki
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 3008 - 3016
  • [4] Multi-task learning for abstractive text summarization with key information guide network
    Weiran Xu
    Chenliang Li
    Minghao Lee
    Chi Zhang
    [J]. EURASIP Journal on Advances in Signal Processing, 2020
  • [5] Multi-task learning for abstractive text summarization with key information guide network
    Xu, Weiran
    Li, Chenliang
    Lee, Minghao
    Zhang, Chi
    [J]. EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2020, 2020 (01)
  • [6] Abstractive Summarization with the Aid of Extractive Summarization
    Chen, Yangbin
    Ma, Yun
    Mao, Xudong
    Li, Qing
    [J]. WEB AND BIG DATA (APWEB-WAIM 2018), PT I, 2018, 10987 : 3 - 15
  • [7] A Dialogues Summarization Algorithm Based on Multi-task Learning
    Chen, Haowei
    Li, Chen
    Liang, Jiajing
    Tian, Lihua
    [J]. NEURAL PROCESSING LETTERS, 2024, 56 (03)
  • [8] Plausibility-promoting generative adversarial network for abstractive text summarization with multi-task constraint
    Yang, Min
    Wang, Xintong
    Lu, Yao
    Lv, Jianming
    Shen, Ying
    Li, Chengming
    [J]. INFORMATION SCIENCES, 2020, 521 : 46 - 61
  • [9] A Combined Extractive With Abstractive Model for Summarization
    Liu, Wenfeng
    Gao, Yaling
    Li, Jinming
    Yang, Yuzhen
    [J]. IEEE ACCESS, 2021, 9 : 43970 - 43980
  • [10] SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization
    Ravaut, Mathieu
    Joty, Shafiq
    Chen, Nancy F.
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4504 - 4524