Multilingual hope speech detection: A Robust framework using transfer learning of fine-tuning RoBERTa model

被引:10
|
作者
Malik, Muhammad Shahid Iqbal [1 ]
Nazarova, Anna [1 ]
Jamjoom, Mona Mamdouh [2 ]
Ignatov, Dmitry I. [1 ]
机构
[1] Natl Res Univ, Higher Sch Econ, Fac Comp Sci, 11 Pokrovsky Blvd, Moscow 109028, Russia
[2] Princess Nourah Bint Abdulrahman Univ, Coll Comp & Informat Sci, Dept Comp Sci, Riyadh, Saudi Arabia
关键词
Transfer learning; Russian; XLM-RoBERTa; Hope speech; Translation-based; Multi; -lingual;
D O I
10.1016/j.jksuci.2023.101736
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hope Speech Detection (HSD) from social media is a new direction for promoting and supporting positive content to encourage harmony and positivity in society. As users of social media belong to different linguistic communities, hope speech detection is rarely studied as a multilingual task considering lowresource languages. Moreover, prior studies explored only monolingual techniques, and the Russian language is not addressed. This study tackles the issue of Multi-lingual Hope Speech Detection (MHSD) in English and Russian languages using the transfer learning paradigm with fine-tuning approach. We explore joint multi-lingual and translation-based approaches to tackle the task of multilingualism, where the latter approach adopts the translation mechanism to transform all content into one language and then classify them. The joint multi-lingual method handles it by designing a universal classifier for various languages. We explore the strengths of the Robustly Optimized BERT Pre-Training Approach (RoBERTa) that showed a benchmark in capturing the semantics and contextual information within the content. The proposed framework consists of several stages: 1) data preprocessing, 2) representation of data using RoBERTa models, 3) fine-tuning phase, and 4) classification of hope speech into two labels. A new Russian corpus for hope speech detection is built, containing YouTube comments. Several experiments are conducted in English and Russian languages by using semi-supervised bilingual English and Russian datasets. The findings show that the proposed framework demonstrated benchmark performance and outperformed the baselines. Furthermore, the translation-based approach (Russian-RoBERTa) offered the best performance by achieving 94% accuracy and 80.24% f1-score. (c) 2023 The Author(s). Published by Elsevier B.V. on behalf of King Saud University. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页数:11
相关论文
共 50 条
  • [1] A multilingual offensive language detection method based on transfer learning from transformer fine-tuning model
    El-Alami, Fatima-zahra
    Alaoui, Said Ouatik El
    Nahnahi, Noureddine En
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (08) : 6048 - 6056
  • [2] Fine-Tuning Transformer Models Using Transfer Learning for Multilingual Threatening Text Identification
    Rehan, Muhammad
    Malik, Muhammad Shahid Iqbal
    Jamjoom, Mona Mamdouh
    IEEE ACCESS, 2023, 11 : 106503 - 106515
  • [3] Multilingual hope speech detection from tweets using transfer learning models
    Ahmad, Muhammad
    Ameer, Iqra
    Sharif, Wareesa
    Usman, Sardar
    Muzamil, Muhammad
    Hamza, Ameer
    Jalal, Muhammad
    Batyrshin, Ildar
    Sidorov, Grigori
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [4] Enhancement of Video Anomaly Detection Performance Using Transfer Learning and Fine-Tuning
    Dilek, Esma
    Dener, Murat
    IEEE ACCESS, 2024, 12 : 73304 - 73322
  • [5] Automatic Cauliflower Disease Detection Using Fine-Tuning Transfer Learning Approach
    Noamaan Abdul Azeem
    Sanjeev Sharma
    Anshul Verma
    SN Computer Science, 5 (7)
  • [6] Transfer Learning Gaussian Anomaly Detection by Fine-tuning Representations
    Rippel, Oliver
    Chavan, Arnav
    Lei, Chucai
    Merhof, Dorit
    IMPROVE: PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND VISION ENGINEERING, 2022, : 45 - 56
  • [7] Transfer Learning With Adaptive Fine-Tuning
    Vrbancic, Grega
    Podgorelec, Vili
    IEEE ACCESS, 2020, 8 (08): : 196197 - 196211
  • [8] Adapting multilingual speech representation model for a new, underresourced language through multilingual fine-tuning and continued pretraining
    Nowakowski, Karol
    Ptaszynski, Michal
    Murasaki, Kyoko
    Nieuwazny, Jagna
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
  • [9] Gemstone classification using ConvNet with transfer learning and fine-tuning
    Freire, Willian M.
    Amaral, Aline M. M. M.
    Costa, Yandre M. G.
    2022 29TH INTERNATIONAL CONFERENCE ON SYSTEMS, SIGNALS AND IMAGE PROCESSING (IWSSIP), 2022,
  • [10] FINE-TUNING TRANSFER LEARNING MODEL IN WOVEN FABRIC PATTERN CLASSIFICATION
    Noprisson H.
    Ermatita E.
    Abdiansah A.
    Ayumi V.
    Purba M.
    Setiawan H.
    International Journal of Innovative Computing, Information and Control, 2022, 18 (06): : 1885 - 1894