Multilayer encoder and single-layer decoder for abstractive Arabic text summarization

被引:11
|
作者
Suleiman, Dima [1 ]
Awajan, Arafat [2 ,3 ]
机构
[1] Univ Jordan, King Abdullah II Sch, Amman, Jordan
[2] Princess Sumaya Univ Technol, Amman, Jordan
[3] Mutah Univ, Al Karak, Jordan
关键词
Deep learning; Abstractive text summarization; Recurrent neural network; Attention mechanism; LSTM; ROUGE;
D O I
10.1016/j.knosys.2021.107791
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, an abstractive Arabic text summarization model that is based on sequence-to-sequence recurrent neural networks is proposed. It consists of a multilayer encoder and single-layer decoder. Encoder layers utilize bidirectional long short-term memory, whereas the decoder utilizes unidirectional long short-term memory. The encoder layers are the input text layer, keywords layer and the name entities layer. Moreover, the decoder uses a global attention mechanism that considers all the input hidden states to generate the summary words. The experiments are conducted on a dataset collected from several resources. The quality of the generated summary is measured quantitatively and qualitatively. In the quantitative measure, in addition to ROUGE1, three new evaluation measures are proposed to evaluate the quality of the generated summary, called ROUGE1-NOORDER, ROUGE1-STEM and ROUGE1-CONTEXT. One of the reasons for proposing new evaluation measures is that the abstractive nature of the summary requires more context based evaluations. Another reason refers to the morphological nature of the Arabic language since several words can be generated from the same root using morphemes. Moreover, a qualitative evaluation measure that is performed by a human is used to evaluate the readability and the relevance of the generated summary since it is hard to automatically measure the readability and relevance. The experimental results show that the multilayer encoder models provide the best results, where the values of ROUGE1, ROUGE1NOORDER, ROUGE1-STEM and ROUGE1-CONTEXT of the proposed model are 38.4, 46.2, 52.6 and 58.1, respectively. Furthermore, the qualitative evaluation shows that the proposed model is the best, achieving an average readability and relevant value equal to 75.9%. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Variational Neural Decoder for Abstractive Text Summarization
    Zhao, Huan
    Cao, Jie
    Xu, Mingquan
    Lu, Jian
    [J]. COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2020, 17 (02) : 537 - 552
  • [2] Keyword-Aware Encoder for Abstractive Text Summarization
    Hu, Tianxiang
    Liang, Jingxi
    Ye, Wei
    Zhang, Shikun
    [J]. DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT II, 2021, 12682 : 37 - 52
  • [3] Abstractive Arabic Text Summarization Based on Deep Learning
    Wazery, Y. M.
    Saleh, Marwa E.
    Alharbi, Abdullah
    Ali, Abdelmgeid A.
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [4] Multi-Encoder Transformer for Korean Abstractive Text Summarization
    Shin, Youhyun
    [J]. IEEE ACCESS, 2023, 11 : 48768 - 48782
  • [5] A Normalized Encoder-Decoder Model for Abstractive Summarization Using Focal Loss
    Shi, Yunsheng
    Meng, Jun
    Wang, Jian
    Lin, Hongfei
    Li, Yumeng
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2018, PT II, 2018, 11109 : 383 - 392
  • [6] A Novel Attention Mechanism considering Decoder Input for Abstractive Text Summarization
    Niu, Jianwei
    Sun, Mingsheng
    Rodrigues, Joel J. P. C.
    Liu, Xuefeng
    [J]. ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [7] English-Arabic Text Translation and Abstractive Summarization Using Transformers
    Holiel, Heidi Ahmed
    Mohamed, Nancy
    Ahmed, Arwa
    Medhat, Walaa
    [J]. 2023 20TH ACS/IEEE INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS, AICCSA, 2023,
  • [8] A Dual Attention Encoder-Decoder Text Summarization Model
    Hakami, Nada Ali
    Mahmoud, Hanan Ahmed Hosni
    [J]. CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 74 (02): : 3697 - 3710
  • [9] SemG-TS: Abstractive Arabic Text Summarization Using Semantic Graph Embedding
    Etaiwi, Wael
    Awajan, Arafat
    [J]. MATHEMATICS, 2022, 10 (18)
  • [10] Abstractive Financial News Summarization via Transformer-BiLSTM Encoder and Graph Attention-Based Decoder
    Li, Haozhou
    Peng, Qinke
    Mou, Xu
    Wang, Ying
    Zeng, Zeyuan
    Bashir, Muhammad Fiaz
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 3190 - 3205