Character Eyes: Seeing Language through Character-Level Taggers

被引:0
|
作者
Pinter, Yuval [1 ]
Marone, Marc [2 ]
Eisenstein, Jacob [3 ]
机构
[1] Georgia Inst Technol, Sch Interact Comp, Atlanta, GA 30332 USA
[2] Microsoft, Redmond, WA USA
[3] Facebook AI Res, Menlo Pk, CA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Character-level models have been used extensively in recent years in NLP tasks as both supplements and replacements for closed-vocabulary token-level word representations. In one popular architecture, character-level LSTMs are used to feed token representations into a sequence tagger predicting token-level annotations such as part-of-speech (POS) tags. In this work, we examine the behavior of POS taggers across languages from the perspective of individual hidden units within the character LSTM. We aggregate the behavior of these units into language-level metrics which quantify the challenges that taggers face on languages with different morphological properties, and identify links between synthesis and affixation preference and emergent behavior of the hidden tagger layer. In a comparative experiment, we show how modifying the balance between forward and backward hidden units affects model arrangement and performance in these types of languages.
引用
收藏
页码:95 / 102
页数:8
相关论文
共 50 条
  • [1] Character-Level Language Modeling with Recurrent Highway Hypernetworks
    Suarez, Joseph
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [2] Character-Level Neural Language Modelling in the Clinical Domain
    Kreuzthaler, Markus
    Oleynik, Michel
    Schulz, Stefan
    [J]. DIGITAL PERSONALIZED HEALTH AND MEDICINE, 2020, 270 : 83 - 87
  • [3] Internal Focalization and Seeing through a Character's Eyes
    Bruhns, Adrian
    Koeppe, Tilmann
    [J]. ESTETIKA-THE EUROPEAN JOURNAL OF AESTHETICS, 2024, 61 (02): : 126 - 141
  • [4] CHARACTER-LEVEL LANGUAGE MODELING WITH HIERARCHICAL RECURRENT NEURAL NETWORKS
    Hwang, Kyuyeon
    Sung, Wonyong
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5720 - 5724
  • [5] Character-Level Language Modeling with Deeper Self-Attention
    Al-Rfou, Rami
    Choe, Dokook
    Constant, Noah
    Guo, Mandy
    Jones, Llion
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 3159 - 3166
  • [6] Character-Level Quantum Mechanical Approach for a Neural Language Model
    Wang, Zhihao
    Ren, Min
    Tian, Xiaoyan
    Liang, Xia
    [J]. INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2019, 12 (02) : 1613 - 1621
  • [7] Character-Level Quantum Mechanical Approach for a Neural Language Model
    Zhihao Wang
    Min Ren
    Xiaoyan Tian
    Xia Liang
    [J]. International Journal of Computational Intelligence Systems, 2019, 12 : 1613 - 1621
  • [8] Parameter-Efficient Korean Character-Level Language Modeling
    Cognetta, Marco
    Moon, Sangwhan
    Wolf-Sonkin, Lawrence
    Okazaki, Naoaki
    [J]. 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2350 - 2356
  • [9] Character-level convolutional networks for arithmetic operator character recognition
    Liang, Zhijie
    Li, Qing
    Liao, Shengbin
    [J]. FIFTH INTERNATIONAL CONFERENCE ON EDUCATIONAL INNOVATION THROUGH TECHNOLOGY (EITT 2016), 2016, : 208 - 212
  • [10] A Character-Level Restoration of Sukhothai Inscriptions Using The Masked Language Model
    Tongkhum, Sujitra
    Sinthupinyo, Sukree
    [J]. 2023 18TH INTERNATIONAL JOINT SYMPOSIUM ON ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE PROCESSING, ISAI-NLP, 2023,