Contextualized Word Representations for Self-Attention Network

被引:0
|
作者
Essam, Mariam [1 ]
Eldawlatly, Seif [1 ]
Abbas, Hazem [1 ]
机构
[1] Ain Shams Univ, Comp & Syst Engn Dept, Cairo, Egypt
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Transfer learning is one approach that could be used to better train deep neural networks. It plays a key role in initializing a network in computer vision applications as opposed to implementing a network from scratch which could he time-consuming. Natural Language Processing (NLP) shares a similar concept of transferring from large-scale data. Recent studies demonstrated that pretrained language models can be used to achieve state-of-the-art results on a multitude of NLP tasks such as sentiment analysis, machine translation and text summarization. In this paper, we demonstrate that a free RNN/CNN self attention model used for sentiment analysis can be improved with 2.53% by using contextualized word representation learned in a language modeling task.
引用
收藏
页码:116 / 121
页数:6
相关论文
共 50 条
  • [31] The Impact of Word Splitting on the Semantic Content of Contextualized Word Representations
    Soler, Aina Gari
    Labeau, Matthieu
    Clavel, Chloe
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 299 - 320
  • [32] Analyzing the Variation Property of Contextualized Word Representations
    Mizuki, Sakae
    Okazaki, Naoaki
    AI 2019: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11919 : 393 - 405
  • [33] Commentary: Attentional control and the self: The Self-Attention Network (SAN)
    Garcia, Adolfo M.
    Huepe, David
    Martinez, David
    Morales, Juan P.
    Huepe, Daniela
    Hurtado, Esteban
    Calvo, Noelia
    Ibanez, Agustin
    FRONTIERS IN PSYCHOLOGY, 2015, 6
  • [34] Spatial self-attention network with self-attention distillation for fine-grained image recognitionx2729;
    Baffour, Adu Asare
    Qin, Zhen
    Wang, Yong
    Qin, Zhiguang
    Choo, Kim-Kwang Raymond
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 81
  • [35] TESANet: Self-attention network for olfactory EEG classification
    Tong, Chengxuan
    Ding, Yi
    Liang, Kevin Lim Jun
    Zhang, Zhuo
    Zhang, Haihong
    Guan, Cuntai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [36] Grassmannian Manifold Self-Attention Network for Signal Classification
    Wang, Rui
    Hu, Chen
    Chen, Ziheng
    Wu, Xiao-Jun
    Song, Xiaoning
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 5099 - 5107
  • [37] Quantum mixed-state self-attention network
    Chen, Fu
    Zhao, Qinglin
    Feng, Li
    Chen, Chuangtao
    Lin, Yangbin
    Lin, Jianhong
    NEURAL NETWORKS, 2025, 185
  • [38] A visual self-attention network for facial expression recognition
    Yu, Naigong
    Bai, Deguo
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [39] A Self-attention Based LSTM Network for Text Classification
    Jing, Ran
    2019 3RD INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2019), 2019, 1207
  • [40] SELF-ATTENTION GENERATIVE ADVERSARIAL NETWORK FOR SPEECH ENHANCEMENT
    Huy Phan
    Nguyen, Huy Le
    Chen, Oliver Y.
    Koch, Philipp
    Duong, Ngoc Q. K.
    McLoughlin, Ian
    Mertins, Alfred
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7103 - 7107