An optimized hybrid deep learning model based on word embeddings and statistical features for extractive summarization

被引:1
|
作者
Wazery, Yaser M. [1 ]
Saleh, Marwa E. [1 ]
Ali, Abdelmgeid A. [1 ]
机构
[1] Minia Univ, Fac Comp & Informat, Al Minya, Egypt
关键词
Extractive text summarization; Deep learning; KerasTuner optimization; Convolutional Neural Network; Feed-Forward Neural Network;
D O I
10.1016/j.jksuci.2023.101614
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Extractive summarization has recently gained significant attention as a classification problem at the sentence level. Most current summarization methods rely on only one way of representing sentences in a document (i.e., extracted features, word embeddings, BERT embeddings). However, classification performance and summary generation quality will be improved if we combine two ways of representing sentences. This paper presents a novel extractive text summarization method based on word embeddings and statistical features of a single document. Each sentence is encoded using a Convolutional Neural Network (CNN) and a Feed-Forward Neural Network (FFNN) based on word embeddings and statistical features. CNN and FFNN outputs are concatenated to classify the sentence using a Multilayer Perceptron (MLP). In addition, hybrid model parameters are optimized by the KerasTuner optimization technique to determine the most efficient hybrid model. The proposed method was evaluated on the standard Newsroom dataset. Experiments show that the proposed method effectively captures the document's semantic and statistical information and outperforms deep learning, machine learning, and state-of-the-art approaches with scores of 78.64, 74.05, and 72.08 for ROUGE-1 ROUGE-2, and ROUGEL, respectively. (c) 2023 The Author(s). Published by Elsevier B.V. on behalf of King Saud University. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Deep learning-based extractive text summarization with word-level attention mechanism
    Mahak Gambhir
    Vishal Gupta
    [J]. Multimedia Tools and Applications, 2022, 81 : 20829 - 20852
  • [2] Deep learning-based extractive text summarization with word-level attention mechanism
    Gambhir, Mahak
    Gupta, Vishal
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (15) : 20829 - 20852
  • [3] Central Embeddings for Extractive Summarization Based on Similarity
    Gutierrez-Hinojosa, Sandra J.
    Calvo, Hiram
    Moreno-Armendariz, Marco A.
    [J]. COMPUTACION Y SISTEMAS, 2019, 23 (03): : 649 - 663
  • [4] Graph Based Extractive News Articles Summarization Approach leveraging Static Word Embeddings
    Barman, Utpal
    Barman, Vishal
    Rahman, Mustafizur
    Choudhury, Nawaz Khan
    [J]. 2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL PERFORMANCE EVALUATION (COMPE-2021), 2021, : 8 - 11
  • [5] A survey of word embeddings based on deep learning
    Shirui Wang
    Wenan Zhou
    Chao Jiang
    [J]. Computing, 2020, 102 : 717 - 740
  • [6] A survey of word embeddings based on deep learning
    Wang, Shirui
    Zhou, Wenan
    Jiang, Chao
    [J]. COMPUTING, 2020, 102 (03) : 717 - 740
  • [7] Deep reinforcement learning for extractive document summarization
    Yao, Kaichun
    Zhang, Libo
    Luo, Tiejian
    Wu, Yanjun
    [J]. NEUROCOMPUTING, 2018, 284 : 52 - 62
  • [8] Extractive Text Summarization using Deep Learning
    Shirwandkar, Nikhil S.
    Kulkarni, Samidha
    [J]. 2018 FOURTH INTERNATIONAL CONFERENCE ON COMPUTING COMMUNICATION CONTROL AND AUTOMATION (ICCUBEA), 2018,
  • [9] Extractive summarization based on word information and sentence position
    Cruz, CM
    Urrea, AM
    [J]. COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, 2005, 3406 : 653 - 656
  • [10] Extractive text summarization using deep learning approach
    Yadav A.K.
    Singh A.
    Dhiman M.
    Vineet
    Kaundal R.
    Verma A.
    Yadav D.
    [J]. International Journal of Information Technology, 2022, 14 (5) : 2407 - 2415