Deep Neural Network Models for Paraphrased Text Classification in the Arabic Language

被引:7
|
作者
Mahmoud, Adnen [1 ,2 ]
Zrigui, Mounir [1 ]
机构
[1] Univ Monastir, Algebra Numbers Theory & Nonlinear Analyzes Lab L, Monastir, Tunisia
[2] Univ Sousse, Higher Inst Comp Sci & Commun Tech, Hammam Sousse, Sousse, Tunisia
关键词
Paraphrase detection; Deep learning; Word embedding; Convolutional neural network; Long short term memory; Arabic corpus construction;
D O I
10.1007/978-3-030-23281-8_1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Paraphrase is the act of reusing original texts without proper citation of the source. Different obfuscation operations can be employed such as addition/deletion of words, synonym substitutions, lexical changes, active to passive switching, etc. This phenomenon dramatically increased because of the progressive advancement of the web and the automatic text editing tools. Recently, deep leaning methods have gained competitive results than traditional methods for Natural Language Processing (NLP). In this context, we consider the problem of Arabic paraphrase detection. We present different deep neural networks like Convolutional Neural Network (CNN) and Long Short Term Memory (LSTM). Our aim is to study the effective of each one in extracting the proper features of sentences without the knowledge of semantic and syntactic structure of Arabic language. For the experiments, we propose an automatic corpus construction seeing the lack of Arabic resources publicly available. Evaluations reveal that LSTM model achieved the higher rate of semantic similarity and outperformed significantly other state-of-the-art methods.
引用
收藏
页码:3 / 16
页数:14
相关论文
共 50 条
  • [21] Deep Convolutional Neural Network for Knowledge-Infused Text Classification
    Malik, Sonika
    Jain, Sarika
    [J]. NEW GENERATION COMPUTING, 2024, 42 (01) : 157 - 176
  • [22] Deep Learning Classification of Biomedical Text using Convolutional Neural Network
    Dollah, Rozilawati
    Sheng, Chew Yi
    Zakaria, Norhawaniah
    Othman, Mohd Shahizan
    Rasib, Abd Wahid
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2019, 10 (08) : 512 - 517
  • [23] Multi-applicable text classification based on deep neural network
    Yang, Jingjing
    Deng, Feng
    Lv, Suhuan
    Wang, Rui
    Guo, Qi
    Kou, Zongchun
    Chen, Shiqiang
    [J]. INTERNATIONAL JOURNAL OF SENSOR NETWORKS, 2022, 40 (04) : 277 - 286
  • [24] Deep Convolutional Neural Network for Knowledge-Infused Text Classification
    Sonika Malik
    Sarika Jain
    [J]. New Generation Computing, 2024, 42 : 157 - 176
  • [25] Arabic Text Classification Using Deep Learning Technics
    Boukil, Samir
    Biniz, Mohamed
    El Adnani, Fatiha
    Cherrat, Loubna
    El Moutaouakkil, Abd Elmaj Id
    [J]. INTERNATIONAL JOURNAL OF GRID AND DISTRIBUTED COMPUTING, 2018, 11 (09): : 103 - 114
  • [26] Arabic Text Diacritization Using Deep Neural Networks
    Fadel, Ali
    Tuffaha, Ibraheem
    Al-Jawarneh, Bara
    Al-Ayyoub, Mahmoud
    [J]. 2019 2ND INTERNATIONAL CONFERENCE ON COMPUTER APPLICATIONS & INFORMATION SECURITY (ICCAIS), 2019,
  • [27] Evaluation of Neural Network Language Models In Handwritten Chinese Text Recognition
    Wu, Yi-Chao
    Yin, Fei
    Liu, Cheng-Lin
    [J]. 2015 13TH IAPR INTERNATIONAL CONFERENCE ON DOCUMENT ANALYSIS AND RECOGNITION (ICDAR), 2015, : 166 - 170
  • [28] Improving handwritten Chinese text recognition using neural network language models and convolutional neural network shape models
    Wu, Yi-Chao
    Yin, Fei
    Liu, Cheng-Lin
    [J]. PATTERN RECOGNITION, 2017, 65 : 251 - 264
  • [29] BERT Models for Arabic Text Classification: A Systematic Review
    Alammary, Ali Saleh
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (11):
  • [30] Determining the Age of the Author of the Text Based on Deep Neural Network Models
    Romanov, Aleksandr Sergeevich
    Kurtukova, Anna Vladimirovna
    Sobolev, Artem Alexandrovich
    Shelupanov, Alexander Alexandrovich
    Fedotova, Anastasia Mikhailovna
    [J]. INFORMATION, 2020, 11 (12) : 1 - 12