Summary-Sentence Level Hierarchical Supervision for Re-Ranking Model of Two-Stage Abstractive Summarization Framework

被引:0
|
作者
Yoo, Eunseok [1 ]
Kim, Gyunyeop [1 ]
Kang, Sangwoo [1 ]
机构
[1] Gachon Univ, Sch Comp, Seongnam Si 1342, South Korea
基金
新加坡国家研究基金会;
关键词
abstractive summarization; text summarization; natural language processing; deep learning;
D O I
10.3390/math12040521
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Fine-tuning a pre-trained sequence-to-sequence-based language model has significantly advanced the field of abstractive summarization. However, the early models of abstractive summarization were limited by the gap between training and inference, and they did not fully utilize the potential of the language model. Recent studies have introduced a two-stage framework that allows the second-stage model to re-rank the candidate summary generated by the first-stage model, to resolve these limitations. In this study, we point out that the supervision method performed in the existing re-ranking model of the two-stage abstractive summarization framework cannot learn detailed and complex information of the data. In addition, we present the problem of positional bias in the existing encoder-decoder-based re-ranking model. To address these two limitations, this study proposes a hierarchical supervision method that jointly performs summary and sentence-level supervision. For sentence-level supervision, we designed two sentence-level loss functions: intra- and inter-intra-sentence ranking losses. Compared to the existing abstractive summarization model, the proposed method exhibited a performance improvement for both the CNN/DM and XSum datasets. The proposed model outperformed the baseline model under a few-shot setting.
引用
收藏
页数:17
相关论文
共 8 条
  • [1] SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization
    Ravaut, Mathieu
    Joty, Shafiq
    Chen, Nancy F.
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4504 - 4524
  • [2] Extractive-Abstractive: A Two-Stage Model for Long Text Summarization
    Liang, Rui
    Li, Jianguo
    Huang, Li
    Lin, Ronghua
    Lai, Yu
    Xiong, Dan
    [J]. COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2021, PT II, 2022, 1492 : 173 - 184
  • [3] Two-stage Discriminative Re-ranking for Large-scale Landmark Retrieval
    Yokoo, Shuhei
    Ozaki, Kohei
    Simo-Serra, Edgar
    Iizuka, Satoshi
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 4363 - 4370
  • [4] Discriminative fine-grained network for vehicle re-identification using two-stage re-ranking
    Qi WANG
    Weidong MIN
    Daojing HE
    Song ZOU
    Tiemei HUANG
    Yu ZHANG
    Ruikang LIU
    [J]. Science China(Information Sciences), 2020, 63 (11) : 147 - 158
  • [5] Discriminative fine-grained network for vehicle re-identification using two-stage re-ranking
    Qi Wang
    Weidong Min
    Daojing He
    Song Zou
    Tiemei Huang
    Yu Zhang
    Ruikang Liu
    [J]. Science China Information Sciences, 2020, 63
  • [6] Discriminative fine-grained network for vehicle re-identification using two-stage re-ranking
    Wang, Qi
    Min, Weidong
    He, Daojing
    Zou, Song
    Huang, Tiemei
    Zhang, Yu
    Liu, Ruikang
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2020, 63 (11)
  • [7] Two-stage hierarchical video summary extraction to match low-level user browsing preferences
    Ferman, AM
    Tekalp, AM
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2003, 5 (02) : 244 - 256
  • [8] Hierarchical and two-stage framework for the paced mixed-model assembly line balancing and sequencing problem considering ergonomic risk
    Lin, Libin
    Wei, Lijun
    Liu, Ting
    Zhang, Hao
    Qin, Peihua
    Leng, Jiewu
    Zhang, Ding
    Liu, Qiang
    [J]. ENGINEERING OPTIMIZATION, 2024, 56 (07) : 1098 - 1121