IgboBERT Models: Building and Training Transformer Models for the Igbo Language

被引:0
|
作者
Chukwuneke, Chiamaka [1 ,2 ]
Ezeani, Ignatius [1 ,2 ]
Rayson, Paul [1 ]
El-Haj, Mahmoud [1 ]
机构
[1] Univ Lancaster, UCREL NLP Grp, Lancaster, England
[2] Nnamdi Azikiwe Univ, Dept Comp Sc, Anambra State, Nigeria
关键词
Igbo; named entity recognition; BERT models; under-resourced; dataset;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This work presents a standard Igbo named entity recognition (IgboNER) dataset as well as the results from training and fine-tuning state-of-the-art transformer IgboNER models. We discuss the process of our dataset creation - data collection and annotation and quality checking. We also present experimental processes involved in building an IgboBERT language model from scratch as well as fine-tuning it along with other non-Igbo pre-trained models for the downstream IgboNER task. Our results show that, although the IgboNER task benefited hugely from fine-tuning large transformer model, fine-tuning a transformer model built from scratch with comparatively little Igbo text data seems to yield quite decent results for the IgboNER task. This work will contribute immensely to IgboNLP in particular as well as the wider African and low-resource NLP efforts.
引用
收藏
页码:5114 / 5122
页数:9
相关论文
共 50 条
  • [21] ACOUSTICALLY DISCRIMINATIVE TRAINING FOR LANGUAGE MODELS
    Kurata, Gakuto
    Itoh, Nobuyasu
    Nishimura, Masafumi
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 4717 - 4720
  • [22] Building sustainable information systems and transformer models on demand
    Asselborn, Thomas
    Melzer, Sylvia
    Schiff, Simon
    Bender, Magnus
    Marwitz, Florian Andreas
    Aljoumani, Said
    Thiemann, Stefan
    Hirschler, Konrad
    Moeller, Ralf
    HUMANITIES & SOCIAL SCIENCES COMMUNICATIONS, 2025, 12 (01):
  • [23] Propositional Reasoning via Neural Transformer Language Models
    Tomasic, Anthony
    Romero, Oscar J.
    Zimmerman, John
    Steinfeld, Aaron
    NEURAL-SYMBOLIC LEARNING AND REASONING, NESY 2022, 2022, : 104 - 119
  • [24] Can Transformer Language Models Predict Psychometric Properties?
    Laverghetta, Antonio, Jr.
    Nighojkar, Animesh
    Mirzakhalov, Jamshidbek
    Licato, John
    10TH CONFERENCE ON LEXICAL AND COMPUTATIONAL SEMANTICS (SEM 2021), 2021, : 12 - 25
  • [25] Improved Hybrid Streaming ASR with Transformer Language Models
    Baquero-Arnal, Pau
    Jorge, Javier
    Gimenez, Adria
    Albert Silvestre-Cerda, Joan
    Iranzo-Sanchez, Javier
    Sanchis, Albert
    Civera, Jorge
    Juan, Alfons
    INTERSPEECH 2020, 2020, : 2127 - 2131
  • [26] A Teaching Language for Building Object Detection Models
    Sultanum, Nicole
    Ghorashi, Soroush
    Meek, Christopher
    Ramos, Gonzalo
    PROCEEDINGS OF THE 2020 ACM DESIGNING INTERACTIVE SYSTEMS CONFERENCE (DIS 2020), 2020, : 1223 - 1234
  • [27] Building social cognitive models of language change
    Hruschka, Daniel J.
    Christiansen, Morten H.
    Blythe, Richard A.
    Croft, William
    Heggarty, Paul
    Mufwene, Salikoko S.
    Pierrehumbert, Janet B.
    Poplack, Shana
    TRENDS IN COGNITIVE SCIENCES, 2009, 13 (11) : 464 - 469
  • [28] Building Language Models for Text with Named Entities
    Parvez, Md Rizwan
    Chakraborty, Saikat
    Ray, Baishakhi
    Chang, Kai-Wei
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 2373 - 2383
  • [29] A Possibilistic Approach for Building Statistical Language Models
    Momtazi, Saeedeh
    Sameti, Hossein
    2009 9TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS DESIGN AND APPLICATIONS, 2009, : 1014 - +
  • [30] Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
    Sartran, Laurent
    Barrett, Samuel
    Kuncoro, Adhiguna
    Stanojevic, Milos
    Blunsom, Phil
    Dyer, Chris
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1423 - 1439