IgboBERT Models: Building and Training Transformer Models for the Igbo Language

被引:0
|
作者
Chukwuneke, Chiamaka [1 ,2 ]
Ezeani, Ignatius [1 ,2 ]
Rayson, Paul [1 ]
El-Haj, Mahmoud [1 ]
机构
[1] Univ Lancaster, UCREL NLP Grp, Lancaster, England
[2] Nnamdi Azikiwe Univ, Dept Comp Sc, Anambra State, Nigeria
关键词
Igbo; named entity recognition; BERT models; under-resourced; dataset;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This work presents a standard Igbo named entity recognition (IgboNER) dataset as well as the results from training and fine-tuning state-of-the-art transformer IgboNER models. We discuss the process of our dataset creation - data collection and annotation and quality checking. We also present experimental processes involved in building an IgboBERT language model from scratch as well as fine-tuning it along with other non-Igbo pre-trained models for the downstream IgboNER task. Our results show that, although the IgboNER task benefited hugely from fine-tuning large transformer model, fine-tuning a transformer model built from scratch with comparatively little Igbo text data seems to yield quite decent results for the IgboNER task. This work will contribute immensely to IgboNLP in particular as well as the wider African and low-resource NLP efforts.
引用
收藏
页码:5114 / 5122
页数:9
相关论文
共 50 条
  • [31] Building a training image with Digital Outcrop Models
    Pickel, A.
    Frechette, J. D.
    Comunian, A.
    Weissmann, G. S.
    JOURNAL OF HYDROLOGY, 2015, 531 : 53 - 61
  • [32] Ensemble of Training Models for Road and Building Segmentation
    Kamiya, Ryosuke
    Sawada, Kyoya
    Hotta, Kazuhiro
    2019 DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2019, : 155 - 160
  • [33] Training Connectionist models for the structured language model
    Xu, P
    Emami, A
    Jelinek, F
    PROCEEDINGS OF THE 2003 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, 2003, : 160 - 167
  • [34] BLENDED LEARNING MODELS IN THE TRAINING OF LANGUAGE STUDENTS
    Makarova, Elena, V
    Ivanova, Ekaterina A.
    PHILOLOGICAL CLASS, 2021, 26 (03): : 222 - 230
  • [35] Training Trajectories of Language Models Across Scales
    Xia, Mengzhou
    Artetxe, Mikel
    Zhou, Chunting
    Lin, Xi Victoria
    Pasunuru, Ramakanth
    Chen, Danqi
    Zettlemoyer, Luke
    Stoyanov, Ves
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 13711 - 13738
  • [36] Regularized Training of Nearest Neighbor Language Models
    Ton, Jean-Francois
    Talbott, Walter
    Zhai, Shuangfei
    Susskind, Joshua
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2022, : 25 - 30
  • [37] Hybrid Alignment Training for Large Language Models
    Wang, Chenglong
    Zhou, Hang
    Chang, Kaiyan
    Li, Bei
    Mu, Yongyu
    Xiao, Tong
    Liu, Tongran
    Zhu, Jingbo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 11389 - 11403
  • [38] Fast Parallel Training of Neural Language Models
    Xiao, Tong
    Zhu, Jingbo
    Liu, Tongran
    Zhang, Chunliang
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4193 - 4199
  • [39] Preparing Lessons for Progressive Training on Language Models
    Pan, Yu
    Yuan, Ye
    Yin, Yichun
    Shi, Jiaxin
    Xu, Zenglin
    Zhang, Ming
    Shang, Lifeng
    Jiang, Xin
    Liu, Qun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18860 - 18868
  • [40] Discriminative training of language models for speech recognition
    Kuo, KHJ
    Fosler-Lussier, E
    Jiang, H
    Lee, CH
    2002 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-IV, PROCEEDINGS, 2002, : 325 - 328