Fine-Tuning Transformer Models Using Transfer Learning for Multilingual Threatening Text Identification

被引:6
|
作者
Rehan, Muhammad [1 ]
Malik, Muhammad Shahid Iqbal [2 ]
Jamjoom, Mona Mamdouh [3 ]
机构
[1] Capital Univ Sci & Technol, Dept Comp Sci, Islamabad 44000, Pakistan
[2] Natl Res Univ Higher Sch Econ, Dept Comp Sci, Moscow 109028, Russia
[3] Princess Nourah Bint Abdulrahman Univ, Coll Comp & Informat Sci, Dept Comp Sci, Riyadh 11671, Saudi Arabia
关键词
Multi-lingual; Urdu; XLM-RoBERTa; threatening text; fine-tunning; MuRIL; LANGUAGE DETECTION;
D O I
10.1109/ACCESS.2023.3320062
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Threatening content detection on social media has recently gained attention. There is very limited work regarding threatening content detection in low-resource languages, especially in Urdu. Furthermore, previous work explored only mono-lingual approaches, and multi-lingual threatening content detection was not studied. This research addressed the task of Multi-lingual Threatening Content Detection (MTCD) in Urdu and English languages by exploiting transfer learning methodology with fine-tuning techniques. To address the multi-lingual task, we investigated two methodologies: 1) Joint multi-lingual, and 2) Joint-translated method. The former approach employs the concept of building a universal classifier for different languages whereas the latter approach applies the translation process to transform the text into one language and then perform classification. We explore the Multilingual Representations for Indian Languages (MuRIL) and Robustly Optimized BERT Pre-Training Approach (RoBERTa) with fine-tuning that already demonstrated state-of-the-art in capturing the contextual and semantic characteristics within the text. For hyper-parameters, manual search and grid search strategies are utilized to find the optimum values. Various experiments are performed on bi-lingual English and Urdu datasets and findings revealed that the proposed methodology outperformed the baselines and showed benchmark performance. The RoBERTa model achieved the highest performance by demonstrating 92% accuracy and 90% macro F1-score with the joint multi-lingual approach.
引用
收藏
页码:106503 / 106515
页数:13
相关论文
共 50 条
  • [1] A transformer fine-tuning strategy for text dialect identification
    Mohammad Ali Humayun
    Hayati Yassin
    Junaid Shuja
    Abdullah Alourani
    Pg Emeroylariffion Abas
    Neural Computing and Applications, 2023, 35 : 6115 - 6124
  • [2] A transformer fine-tuning strategy for text dialect identification
    Humayun, Mohammad Ali
    Yassin, Hayati
    Shuja, Junaid
    Alourani, Abdullah
    Abas, Pg Emeroylariffion
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (08): : 6115 - 6124
  • [3] A multilingual offensive language detection method based on transfer learning from transformer fine-tuning model
    El-Alami, Fatima-zahra
    Alaoui, Said Ouatik El
    Nahnahi, Noureddine En
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (08) : 6048 - 6056
  • [4] Knee Implant Identification by Fine-Tuning Deep Learning Models
    Sukkrit Sharma
    Vineet Batta
    Malathy Chidambaranathan
    Prabhakaran Mathialagan
    Gayathri Mani
    M. Kiruthika
    Barun Datta
    Srinath Kamineni
    Guruva Reddy
    Suhas Masilamani
    Sandeep Vijayan
    Derek F. Amanatullah
    Indian Journal of Orthopaedics, 2021, 55 : 1295 - 1305
  • [5] Knee Implant Identification by Fine-Tuning Deep Learning Models
    Sharma, Sukkrit
    Batta, Vineet
    Chidambaranathan, Malathy
    Mathialagan, Prabhakaran
    Mani, Gayathri
    Kiruthika, M.
    Datta, Barun
    Kamineni, Srinath
    Reddy, Guruva
    Masilamani, Suhas
    Vijayan, Sandeep
    Amanatullah, Derek F.
    INDIAN JOURNAL OF ORTHOPAEDICS, 2021, 55 (05) : 1295 - 1305
  • [6] Multilingual hope speech detection: A Robust framework using transfer learning of fine-tuning RoBERTa model
    Malik, Muhammad Shahid Iqbal
    Nazarova, Anna
    Jamjoom, Mona Mamdouh
    Ignatov, Dmitry I.
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (08)
  • [7] Rebetiko Singer Identification: Fine-tuning and explaining deep pretrained transformer models
    Papakostas, Maximos Kaliakatsos
    Zacharakis, Asterios
    Velenis, Konstantinos
    Cambouropoulos, Emilios
    PROCEEDINGS OF THE 19TH INTERNATIONAL AUDIO MOSTLY CONFERENCE, AM 2024, 2024, : 285 - 291
  • [8] Transfer Learning With Adaptive Fine-Tuning
    Vrbancic, Grega
    Podgorelec, Vili
    IEEE ACCESS, 2020, 8 (08): : 196197 - 196211
  • [9] Fine-Tuning QurSim on Monolingual and Multilingual Models for Semantic Search
    Afzal, Tania
    Rauf, Sadaf Abdul
    Malik, Muhammad Ghulam Abbas
    Imran, Muhammad
    INFORMATION, 2025, 16 (02)
  • [10] On fine-tuning deep learning models using transfer learning and hyper-parameters optimization for disease identification in maize leaves
    Malliga Subramanian
    Kogilavani Shanmugavadivel
    P. S. Nandhini
    Neural Computing and Applications, 2022, 34 : 13951 - 13968