Exploiting deep representations for natural language processing

被引:5
|
作者
Zi-Yi Dou [1 ]
Xing Wang [2 ]
Shuming Shi [3 ]
Zhaopeng Tu [2 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[2] Tencent AI Lab, Shenzhen, Peoples R China
[3] Tencent AI Lab, Nat Language Proc Ctr, Shenzhen, Peoples R China
关键词
Natural language processing; Deep neural networks; Deep representations; Layer aggregation; Routing-by-agreement;
D O I
10.1016/j.neucom.2019.12.060
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Advanced neural network models generally implement systems as multiple layers to model complex functions and capture complicated linguistic structures at different levels [1]. However, only the top layers of deep networks are leveraged in the subsequent process, which misses the opportunity to exploit the useful information embedded in other layers. In this work, we propose to expose all of these embedded signals with two types of mechanisms, namely deep connections and iterative routings. While deep connections allow better information and gradient flow across layers, iterative routings directly combine the layer representations to form a final output with iterative routing-by-agreement mechanism. Experimental results on both machine translation and language representation tasks demonstrate the effectiveness and universality of the proposed approaches, which indicates the necessity of exploiting deep representations for natural language processing tasks. While the two strategies individually boost performance, combining them can further improve performance. (c) 2019ElsevierB.V. Allrightsreserved.
引用
收藏
页码:1 / 7
页数:7
相关论文
共 50 条
  • [1] Semantic Representations for Multilingual Natural Language Processing
    Kozerenko, Elena B.
    [J]. 2019 6TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2019), 2019, : 433 - 438
  • [2] Symbolic, Distributed, and Distributional Representations for Natural Language Processing in the Era of Deep Learning: A Survey
    Ferrone, Lorenzo
    Zanzotto, Fabio Massimo
    [J]. FRONTIERS IN ROBOTICS AND AI, 2020, 6
  • [3] Deep Learning in Natural Language Processing
    Feng, Haoda
    Shi, Feng
    [J]. NATURAL LANGUAGE ENGINEERING, 2021, 27 (03) : 373 - 375
  • [4] Deep learning of the natural language processing
    Allauzen, Alexandre
    Schuetze, Hinrich
    [J]. TRAITEMENT AUTOMATIQUE DES LANGUES, 2018, 59 (02): : 7 - 14
  • [5] Deep Learning for Natural Language Processing and Language Modelling
    Klosowski, Piotr
    [J]. 2018 SIGNAL PROCESSING: ALGORITHMS, ARCHITECTURES, ARRANGEMENTS, AND APPLICATIONS (SPA), 2018, : 223 - 228
  • [7] Federated Learning for Exploiting Annotators' Disagreements in Natural Language Processing
    Rodriguez-Barroso, Nuria
    Camara, Eugenio Martinez
    Collados, Jose Camacho
    Luzon, M. Victoria
    Herrera, Francisco
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 630 - 648
  • [8] Deep Learning Methods in Natural Language Processing
    Flores, Alexis Stalin Alulema
    [J]. APPLIED TECHNOLOGIES (ICAT 2019), PT II, 2020, 1194 : 92 - 107
  • [9] Deep Learning on Graphs for Natural Language Processing
    Wu, Lingfei
    Chen, Yu
    Ji, Heng
    Liu, Bang
    [J]. KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 4084 - 4085
  • [10] Deep Natural Language Processing for Search Systems
    Guo, Weiwei
    Gao, Huiji
    Shi, Jun
    Long, Bo
    [J]. PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 1405 - 1406