End-to-End Transformer-Based Models in Textual-Based NLP

被引:27
|
作者
Rahali, Abir [1 ]
Akhloufi, Moulay A. [1 ]
机构
[1] Univ Moncton, Dept Comp Sci, Percept Robot & Intelligent Machines Res Grp PRIME, Moncton, NB E1A 3E9, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Transformers; deep learning; natural language processing; transfer learning; PRE-TRAINED BERT; PREDICTION; SYSTEMS;
D O I
10.3390/ai4010004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer architectures are highly expressive because they use self-attention mechanisms to encode long-range dependencies in the input sequences. In this paper, we present a literature review on Transformer-based (TB) models, providing a detailed overview of each model in comparison to the Transformer's standard architecture. This survey focuses on TB models used in the field of Natural Language Processing (NLP) for textual-based tasks. We begin with an overview of the fundamental concepts at the heart of the success of these models. Then, we classify them based on their architecture and training mode. We compare the advantages and disadvantages of popular techniques in terms of architectural design and experimental value. Finally, we discuss open research, directions, and potential future work to help solve current TB application challenges in NLP.
引用
收藏
页码:54 / 110
页数:57
相关论文
共 50 条
  • [41] Multi-Encoder Learning and Stream Fusion for Transformer-Based End-to-End Automatic Speech Recognition
    Lohrenz, Timo
    Li, Zhengyang
    Fingscheidt, Tim
    INTERSPEECH 2021, 2021, : 2846 - 2850
  • [42] End-to-End Asbestos Roof Detection on Orthophotos Using Transformer-Based YOLO Deep Neural Network
    Pace, Cesare Davide
    Bria, Alessandro
    Focareta, Mariano
    Lozupone, Gabriele
    Marrocco, Claudio
    Meoli, Giuseppe
    Molinara, Mario
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2023, PT I, 2023, 14233 : 232 - 244
  • [43] TLLFusion: An End-to-End Transformer-Based Method for Low-Light Infrared and Visible Image Fusion
    Lv, Guohua
    Fu, Xinyue
    Zhai, Yi
    Zhao, Guixin
    Gao, Yongbiao
    PATTERN RECOGNITION AND COMPUTER VISION, PT III, PRCV 2024, 2025, 15033 : 364 - 378
  • [44] Improving Transformer-based End-to-End Speech Recognition with Connectionist Temporal Classification and Language Model Integration
    Karita, Shigeki
    Soplin, Nelson Enrique Yalta
    Watanabe, Shinji
    Delcroix, Marc
    Ogawa, Atsunori
    Nakatani, Tomohiro
    INTERSPEECH 2019, 2019, : 1408 - 1412
  • [45] End-to-End lightweight Transformer-Based neural network for grasp detection towards fruit robotic handling
    Guo, Congmin
    Zhu, Chenhao
    Liu, Yuchen
    Huang, Renjun
    Cao, Boyuan
    Zhu, Qingzhen
    Zhang, Ranxin
    Zhang, Baohua
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 221
  • [46] CLTR: An End-to-End, Transformer-Based System for Cell Level Table Retrieval and Table Question Answering
    Pan, Feifei
    Canim, Mustafa
    Glass, Michael
    Gliozzo, Alfio
    Fox, Peter
    ACL-IJCNLP 2021: THE JOINT CONFERENCE OF THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE SYSTEM DEMONSTRATIONS, 2021, : 202 - 209
  • [47] Fast offline transformer-based end-to-end automatic speech recognition for real-world applications
    Oh, Yoo Rhee
    Park, Kiyoung
    Park, Jeon Gue
    ETRI JOURNAL, 2022, 44 (03) : 476 - 490
  • [48] CarcassFormer: an end-to-end transformer-based framework for simultaneous localization, segmentation and classification of poultry carcass defect
    Tran, Minh
    Truong, Sang
    Fernandes, Arthur F. A.
    Kidd, Michael T.
    Le, Ngan
    POULTRY SCIENCE, 2024, 103 (08)
  • [49] End to end transformer-based contextual speech recognition based on pointer network
    Lin, Binghuai
    Wang, Liyuan
    INTERSPEECH 2021, 2021, : 2087 - 2091
  • [50] Hierarchical transformer-based large-context end-to-end ASR with large-context knowledge distillation
    Masumura, Ryo
    Makishima, Naoki
    Ihori, Mana
    Takashima, Akihiko
    Tanaka, Tomohiro
    Orihashi, Shota
    arXiv, 2021,