Multi-task Projected Embedding for Igbo

被引:0
|
作者
Ezeani, Ignatius [1 ]
Hepple, Mark [1 ]
Onyenwe, Ikechukwu [1 ]
Enemuo, Chioma [1 ]
机构
[1] Univ Sheffield, Dept Comp Sci, Sheffield, S Yorkshire, England
来源
关键词
Low-resource; Igbo; Diacritics; Embedding models; Transfer learning;
D O I
10.1007/978-3-030-00794-2_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
NLP research on low resource African languages is often impeded by the unavailability of basic resources: tools, techniques, annotated corpora, and datasets. Besides the lack of funding for the manual development of these resources, building from scratch will amount to the reinvention of the wheel. Therefore, adapting existing techniques and models from well-resourced languages is often an attractive option. One of the most generally applied NLP models is word embeddings. Embedding models often require large amounts of data to train which are not available for most African languages. In this work, we adopt an alignment based projection method to transfer trained English embeddings to the Igbo language. Various English embedding models were projected and evaluated on the odd-word, analogy and word-similarity tasks intrinsically, and also on the diacritic restoration task. Our results show that the projected embeddings performed very well across these tasks.
引用
收藏
页码:285 / 294
页数:10
相关论文
共 50 条
  • [1] Multi-task Network Embedding
    Xu, Linchuan
    Wei, Xiaokai
    Cao, Jiannong
    Yu, Philip S.
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2017, : 571 - 580
  • [2] Multi-task network embedding
    Linchuan Xu
    Xiaokai Wei
    Jiannong Cao
    Philip S. Yu
    [J]. International Journal of Data Science and Analytics, 2019, 8 : 183 - 198
  • [3] Multi-task network embedding
    Xu, Linchuan
    Wei, Xiaokai
    Cao, Jiannong
    Yu, Philip S.
    [J]. INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2019, 8 (02) : 183 - 198
  • [4] Task-Projected Hyperdimensional Computing for Multi-task Learning
    Chang, Cheng-Yang
    Chuang, Yu-Chuan
    Wu, An-Yeu
    [J]. ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2020, PT I, 2020, 583 : 241 - 251
  • [5] Multi-Task Label Embedding for Text Classification
    Zhang, Honglun
    Xiao, Liqiang
    Chen, Wenqing
    Wang, Yongkun
    Jin, Yaohui
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4545 - 4553
  • [6] Multi-Task Learning Based Network Embedding
    Wang, Shanfeng
    Wang, Qixiang
    Gong, Maoguo
    [J]. FRONTIERS IN NEUROSCIENCE, 2020, 13
  • [7] Unified Voice Embedding through Multi-task Learning
    Rajenthiran, Jenarthanan
    Sithamaparanathan, Lakshikka
    Uthayakumar, Saranya
    Thayasivam, Uthayasanker
    [J]. 2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 178 - 183
  • [8] Attributed graph clustering with multi-task embedding learning
    Zhang, Xiaotong
    Liu, Han
    Zhang, Xianchao
    Liu, Xinyue
    [J]. NEURAL NETWORKS, 2022, 152 : 224 - 233
  • [9] Multi-task Network Embedding with Adaptive Loss Weighting
    Rizi, Fatemeh Salehi
    Granitzer, Michael
    [J]. 2020 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING (ASONAM), 2020, : 1 - 5
  • [10] MultiE: Multi-Task Embedding for Knowledge Base Completion
    Zhang, Zhao
    Zhuang, Fuzhen
    Niu, Zheng-Yu
    Wang, Deqing
    He, Qing
    [J]. CIKM'18: PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2018, : 1715 - 1718