IgboBERT Models: Building and Training Transformer Models for the Igbo Language

被引:0
|
作者
Chukwuneke, Chiamaka [1 ,2 ]
Ezeani, Ignatius [1 ,2 ]
Rayson, Paul [1 ]
El-Haj, Mahmoud [1 ]
机构
[1] Univ Lancaster, UCREL NLP Grp, Lancaster, England
[2] Nnamdi Azikiwe Univ, Dept Comp Sc, Anambra State, Nigeria
关键词
Igbo; named entity recognition; BERT models; under-resourced; dataset;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This work presents a standard Igbo named entity recognition (IgboNER) dataset as well as the results from training and fine-tuning state-of-the-art transformer IgboNER models. We discuss the process of our dataset creation - data collection and annotation and quality checking. We also present experimental processes involved in building an IgboBERT language model from scratch as well as fine-tuning it along with other non-Igbo pre-trained models for the downstream IgboNER task. Our results show that, although the IgboNER task benefited hugely from fine-tuning large transformer model, fine-tuning a transformer model built from scratch with comparatively little Igbo text data seems to yield quite decent results for the IgboNER task. This work will contribute immensely to IgboNLP in particular as well as the wider African and low-resource NLP efforts.
引用
收藏
页码:5114 / 5122
页数:9
相关论文
共 50 条
  • [41] Probabilistic generative transformer language models for generative design of molecules
    Wei, Lai
    Fu, Nihang
    Song, Yuqi
    Wang, Qian
    Hu, Jianjun
    JOURNAL OF CHEMINFORMATICS, 2023, 15 (01)
  • [42] MIXED PRECISION QUANTIZATION OF TRANSFORMER LANGUAGE MODELS FOR SPEECH RECOGNITION
    Xu, Junhao
    Hu, Shoukang
    Yu, Jianwei
    Liu, Xunying
    Meng, Helen
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7383 - 7387
  • [43] Transformer Language Models Handle Word Frequency in Prediction Head
    Kobayashi, Goro
    Kuribayashi, Tatsuki
    Yokoi, Sho
    Inui, Kentaro
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4523 - 4535
  • [44] ON THE PROTO-ROLE PROPERTIES INFERRED BY TRANSFORMER LANGUAGE MODELS
    Proietti, Mattia
    Lebani, Gianluca e.
    Lenci, Alessandro
    LINGUE E LINGUAGGIO, 2024, 23 (01) : 111 - 140
  • [45] Transformer-Based Language Models for Software Vulnerability Detection
    Thapa, Chandra
    Jang, Seung Ick
    Ahmed, Muhammad Ejaz
    Camtepe, Seyit
    Pieprzyk, Josef
    Nepal, Surya
    PROCEEDINGS OF THE 38TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2022, 2022, : 481 - 496
  • [46] A Comparison of Transformer-Based Language Models on NLP Benchmarks
    Greco, Candida Maria
    Tagarelli, Andrea
    Zumpano, Ester
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 490 - 501
  • [47] Pushdown Layers: Encoding Recursive Structure in Transformer Language Models
    Murty, Shikhar
    Sharma, Pratyusha
    Andreas, Jacob
    Manning, Christopher D.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3233 - 3247
  • [48] Applications of transformer-based language models in bioinformatics: a survey
    Zhang, Shuang
    Fan, Rui
    Liu, Yuti
    Chen, Shuang
    Liu, Qiao
    Zeng, Wanwen
    NEURO-ONCOLOGY ADVANCES, 2023, 5 (01)
  • [49] TAG: Gradient Attack on Transformer-based Language Models
    Deng, Jieren
    Wang, Yijue
    Li, Ji
    Wang, Chenghong
    Shang, Chao
    Liu, Hang
    Rajasekaran, Sanguthevar
    Ding, Caiwen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3600 - 3610
  • [50] On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
    Pilault, Jonathan
    Li, Raymond
    Subramanian, Sandeep
    Pal, Christopher
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9308 - 9319