Contextualized Word Representations for Self-Attention Network

被引:0
|
作者
Essam, Mariam [1 ]
Eldawlatly, Seif [1 ]
Abbas, Hazem [1 ]
机构
[1] Ain Shams Univ, Comp & Syst Engn Dept, Cairo, Egypt
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Transfer learning is one approach that could be used to better train deep neural networks. It plays a key role in initializing a network in computer vision applications as opposed to implementing a network from scratch which could he time-consuming. Natural Language Processing (NLP) shares a similar concept of transferring from large-scale data. Recent studies demonstrated that pretrained language models can be used to achieve state-of-the-art results on a multitude of NLP tasks such as sentiment analysis, machine translation and text summarization. In this paper, we demonstrate that a free RNN/CNN self attention model used for sentiment analysis can be improved with 2.53% by using contextualized word representation learned in a language modeling task.
引用
收藏
页码:116 / 121
页数:6
相关论文
共 50 条
  • [1] LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention
    Yamada, Ikuya
    Asai, Akari
    Shindo, Hiroyuki
    Takeda, Hideaki
    Matsumoto, Yuji
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 6442 - 6454
  • [2] Graph contextualized self-attention network for software service sequential recommendation
    Fu, Zixuan
    Wang, Chenghua
    Xu, Jiajie
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 149 : 509 - 517
  • [3] Graph contextualized self-attention network for session-based recommendation
    Xu, Chengfeng
    Zhao, Pengpeng
    Liu, Yanchi
    Sheng, Victor S.
    Xu, Jiajie
    Zhuang, Fuzhen
    Fang, Junhua
    Zhou, Xiaofang
    IJCAI International Joint Conference on Artificial Intelligence, 2019, 2019-August : 3940 - 3946
  • [4] Graph Contextualized Self-Attention Network for Session-based Recommendation
    Xu, Chengfeng
    Zhao, Pengpeng
    Liu, Yanchi
    Sheng, Victor S.
    Xu, Jiajie
    Zhuang, Fuzhen
    Fang, Junhua
    Zhou, Xiaofang
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3940 - 3946
  • [5] Investigating Self-Attention Network for Chinese Word Segmentation
    Gan, Leilei
    Zhang, Yue
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 2933 - 2941
  • [6] Self-Attention with Structural Position Representations
    Wang, Xing
    Tu, Zhaopeng
    Wang, Longyue
    Shi, Shuming
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1403 - 1409
  • [7] CRSANet: Class Representations Self-Attention network for the segmentation of thyroid nodules
    Sun, Shiyao
    Fu, Chong
    Xu, Sen
    Wen, Yingyou
    Ma, Tao
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 91
  • [8] The function of the self-attention network
    Cunningham, Sheila J.
    COGNITIVE NEUROSCIENCE, 2016, 7 (1-4) : 21 - 22
  • [9] Contextualized Diachronic Word Representations
    Jawahar, Ganesh
    Seddah, Djame
    1ST INTERNATIONAL WORKSHOP ON COMPUTATIONAL APPROACHES TO HISTORICAL LANGUAGE CHANGE, 2019, : 35 - 47
  • [10] Recasting Self-Attention with Holographic Reduced Representations
    Alam, Mohammad Mahmudul
    Raff, Edward
    Biderman, Stella
    Oates, Tim
    Holt, James
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202 : 490 - 507