Text Simplification Using Transformer and BERT

被引:1
|
作者
Alissa, Sarah [1 ]
Wald, Mike [2 ]
机构
[1] Imam Abdulrahman Bin Faisal Univ, Coll Comp Sci & Informat Technol, Dammam, Saudi Arabia
[2] Univ Southampton, Sch Elect & Comp Sci, Southampton, England
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2023年 / 75卷 / 02期
关键词
Text simplification; neural machine translation; transformer;
D O I
10.32604/cmc.2023.033647
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reading and writing are the main interaction methods with web content. Text simplification tools are helpful for people with cognitive impairments, new language learners, and children as they might find difficulties in understanding the complex web content. Text simplification is the process of changing complex text into more readable and understandable text. The recent approaches to text simplification adopted the machine translation concept to learn simplification rules from a parallel corpus of complex and simple sentences. In this paper, we propose two models based on the transformer which is an encoder-decoder structure that achieves state-of-the-art (SOTA) results in machine translation. The training process for our model includes three steps: preprocessing the data using a subword tokenizer, training the model and optimizing it using the Adam optimizer, then using the model to decode the output. The first model uses the transformer only and the second model uses and integrates the Bidirectional Encoder Representations from Transformer (BERT) as encoder to enhance the training time and results. The performance of the proposed model using the transformer was evaluated using the Bilingual Evaluation Understudy score (BLEU) and recorded (53.78) on the WikiSmall dataset. On the other hand, the experiment on the second model which is integrated with BERT shows that the validation loss decreased very fast compared with the model without the BERT. However, the BLEU score was small (44.54), which could be due to the size of the dataset so the model was overfitting and unable to generalize well. Therefore, in the future, the second model could involve experimenting with a larger dataset such as the WikiLarge. In addition, more analysis has been done on the model's results and the used dataset using different evaluation metrics to understand their performance.
引用
收藏
页码:3479 / 3495
页数:17
相关论文
共 50 条
  • [21] VD-BERT: A Unified Vision and Dialog Transformer with BERT
    Wang, Yue
    Joty, Shafiq
    Lyu, Michael R.
    King, Irwin
    Xiong, Caiming
    Hoi, Steven C. H.
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3325 - 3338
  • [22] Effective text classification using BERT, MTM LSTM, and DT
    Jamshidi, Saman
    Mohammadi, Mahin
    Bagheri, Saeed
    Najafabadi, Hamid Esmaeili
    Rezvanian, Alireza
    Gheisari, Mehdi
    Ghaderzadeh, Mustafa
    Shahabi, Amir Shahab
    Wu, Zongda
    DATA & KNOWLEDGE ENGINEERING, 2024, 151
  • [23] Traffic Incident Duration Prediction using BERT Representation of Text
    Agrawal, Prashansa
    Franklin, Antony
    Pawar, Digvijay
    Srijith, P. K.
    2021 IEEE 94TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2021-FALL), 2021,
  • [24] Using BERT and Knowledge Graph for detecting triples in Vietnamese text
    Phuc Do
    Hung Le
    An B. Pham
    Cuong H. Nguyen
    Neural Computing and Applications, 2022, 34 : 17999 - 18013
  • [25] Text Classification of Modern Mongolian Documents using BERT models
    Khaltarkhuu, Garmaabazar
    Batjargal, Biligsaikhan
    Maeda, Akira
    2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 219 - 224
  • [26] Performance Study on Extractive Text Summarization Using BERT Models
    Abdel-Salam, Shehab
    Rafea, Ahmed
    INFORMATION, 2022, 13 (02)
  • [27] Using BERT and Knowledge Graph for detecting triples in Vietnamese text
    Phuc Do
    Hung Le
    Pham, An B.
    Nguyen, Cuong H.
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (20): : 17999 - 18013
  • [28] Turkish abstractive text document summarization using text to text transfer transformer
    Ay, Betul
    Ertam, Fatih
    Fidan, Guven
    Aydin, Galip
    ALEXANDRIA ENGINEERING JOURNAL, 2023, 68 : 1 - 13
  • [29] Text Simplification Tools: Using Machine Learning to Discover Features that Identify Difficult Text
    Kauchak, David
    Mouradi, Obay
    Pentoney, Christopher
    Leroy, Gondy
    2014 47TH HAWAII INTERNATIONAL CONFERENCE ON SYSTEM SCIENCES (HICSS), 2014, : 2616 - 2625
  • [30] Fake News Classification using transformer based enhanced LSTM and BERT
    Rai N.
    Kumar D.
    Kaushik N.
    Raj C.
    Ali A.
    International Journal of Cognitive Computing in Engineering, 2022, 3 : 98 - 105