Techniques Comparison for Natural Language Processing

被引:0
|
作者
Iosifova, Olena [1 ]
Iosifov, Ievgen [1 ]
Rolik, Oleksandr [2 ]
Sokolov, Volodymyr [3 ]
机构
[1] Ender Turing OU, Tallinn, Estonia
[2] Natl Tech Univ Ukraine, Igor Sikorsky Kyiv Polytech Inst, Kiev, Ukraine
[3] Borys Grinchenko Kyiv Univ, Kiev, Ukraine
关键词
Natural Language Processing; NLP; Language Model; Embedding; Recurrent Neural Network; RNN; Gated Recurrent Unit; GRU; Long Short-Term Memory; LSTM; Encoder; Decoder; Attention; Transformer; Transfer Learning; Deep Learning; Neural Network;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
These improvements open many possibilities in solving Natural Language Processing downstream tasks. Such tasks include machine translation, speech recognition, information retrieval, sentiment analysis, summarization, question answering, multilingual dialogue systems development, and many more. Language models are one of the most important components in solving each of the mentioned tasks. This paper is devoted to research and analysis of the most adopted techniques and designs for building and training language models that show a state of the art results. Techniques and components applied in the creation of language models and its parts are observed in this paper, paying attention to neural networks, embedding mechanisms, bidirectionality, encoder and decoder architecture, attention, and self-attention, as well as parallelization through using transformer. As a result, the most promising techniques imply pre-training and fine-tuning of a language model, attention-based neural network as a part of model design, and a complex ensemble of multidimensional embedding to build deep context understanding. The latest offered architectures based on these approaches require a lot of computational power for training language models, and it is a direction of further improvement. Algorithm for choosing right model for relevant business task provided considering current challenges and available architectures.
引用
收藏
页数:11
相关论文
共 50 条
  • [2] Data augmentation techniques in natural language processing
    Pellicer, Lucas Francisco Amaral Orosco
    Ferreira, Taynan Maier
    Costa, Anna Helena Reali
    [J]. APPLIED SOFT COMPUTING, 2023, 132
  • [3] Deep Learning Techniques for Natural Language Processing
    Rodzin, Sergey
    Bova, Victoria
    Kravchenko, Yury
    Rodzina, Lada
    [J]. ARTIFICIAL INTELLIGENCE TRENDS IN SYSTEMS, VOL 2, 2022, 502 : 121 - 130
  • [4] Survey of Natural Language Processing Techniques in Bioinformatics
    Zeng, Zhiqiang
    Shi, Hua
    Wu, Yun
    Hong, Zhiling
    [J]. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2015, 2015
  • [5] Comparison of large language models and traditional natural language processing techniques in predicting arteriovenous fistula failure
    Lama, Suman
    Zhang, Hanjie
    Monaghan, Caitlin
    Bellocchio, Francesco
    Chaudhuri, Sheetal
    Neri, Luca
    Usvyat, Len
    [J]. NEPHROLOGY DIALYSIS TRANSPLANTATION, 2024, 39 : I1303 - I1304
  • [6] Text Classification for Clinical Trial Operations: Evaluation and Comparison of Natural Language Processing Techniques
    Richard, Emma
    Reddy, Bhargava
    [J]. THERAPEUTIC INNOVATION & REGULATORY SCIENCE, 2021, 55 (02) : 447 - 453
  • [7] Text Classification for Clinical Trial Operations: Evaluation and Comparison of Natural Language Processing Techniques
    Emma Richard
    Bhargava Reddy
    [J]. Therapeutic Innovation & Regulatory Science, 2021, 55 : 447 - 453
  • [8] Symbiosis of evolutionary techniques and statistical natural language processing
    Araujo, L
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2004, 8 (01) : 14 - 27
  • [9] Defining the Malice Space with Natural Language Processing Techniques
    Patten, Terry
    Call, Catherine
    Mitchell, Daniel
    Taylor, Jason
    Lasser, Samuel
    [J]. 2016 CYBERSECURITY SYMPOSIUM, 2016, : 44 - 50
  • [10] Processing natural language without natural language processing
    Brill, E
    [J]. COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, PROCEEDINGS, 2003, 2588 : 360 - 369