Bidirectional Language Modeling: A Systematic Literature Review

被引:4
|
作者
Shah Jahan, Muhammad [1 ]
Khan, Habib Ullah [2 ]
Akbar, Shahzad [3 ]
Umar Farooq, Muhammad [1 ]
Gul, Sarah [4 ]
Amjad, Anam [1 ]
机构
[1] Natl Univ Sci & Technol, Dept Comp Engn, Coll Elect & Mech Engn, Islamabad 44000, Pakistan
[2] Qatar Univ, Dept Accounting & Informat Syst, Coll Business & Econ, Doha, Qatar
[3] Riphah Int Univ, Riphah Coll Comp, Faisalabad Campus, Faisalabad 3800, Pakistan
[4] Int Islamic Univ, Dept Biol Sci, FBAS, Islamabad, Pakistan
关键词
All Open Access; Gold;
D O I
10.1155/2021/6641832
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In transfer learning, two major activities, i.e., pretraining and fine-tuning, are carried out to perform downstream tasks. The advent of transformer architecture and bidirectional language models, e.g., bidirectional encoder representation from transformer (BERT), enables the functionality of transfer learning. Besides, BERT bridges the limitations of unidirectional language models by removing the dependency on the recurrent neural network (RNN). BERT also supports the attention mechanism to read input from any side and understand sentence context better. It is analyzed that the performance of downstream tasks in transfer learning depends upon the various factors such as dataset size, step size, and the number of selected parameters. In state-of-the-art, various research studies produced efficient results by contributing to the pretraining phase. However, a comprehensive investigation and analysis of these research studies is not available yet. Therefore, in this article, a systematic literature review (SLR) is presented investigating thirty-one (31) influential research studies published during 2018-2020. Following contributions are made in this paper: (1) thirty-one (31) models inspired by BERT are extracted. (2) Every model in this paper is compared with RoBERTa (replicated BERT model) having large dataset and batch size but with a small step size. It is concluded that seven (7) out of thirty-one (31) models in this SLR outperforms RoBERTa in which three were trained on a larger dataset while the other four models are trained on a smaller dataset. Besides, among these seven models, six models shared both feedforward network (FFN) and attention across the layers. Rest of the twenty-four (24) models are also studied in this SLR with different parameter settings. Furthermore, it has been concluded that a pretrained model with a large dataset, hidden layers, attention heads, and small step size with parameter sharing produces better results. This SLR will help researchers to pick a suitable model based on their requirements.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] A Systematic Literature Review on Interaction Flow Modeling Language (IFML)
    Hamdani, Maryum
    Butt, Wasi Haider
    Anwar, Muhammad Waseem
    Azam, Farooque
    [J]. PROCEEDINGS OF THE 2018 2ND INTERNATIONAL CONFERENCE ON MANAGEMENT ENGINEERING, SOFTWARE ENGINEERING AND SERVICE SCIENCES (ICMSS 2018), 2018, : 134 - 138
  • [2] Natural Language Processing in Business Process Identification and Modeling: A Systematic Literature Review
    de Almeida Bordignon, Ana Claudia
    Thom, Lucineia Heloisa
    Silva, Thanner Soares
    Dani, Vinicius Stein
    Fantinato, Marcelo
    Borges Ferreira, Renato Cesar
    [J]. PROCEEDINGS OF THE 14TH BRAZILIAN SYMPOSIUM ON INFORMATION SYSTEMS (SBSI2018), 2018, : 191 - 198
  • [3] Threat modeling - A systematic literature review
    Xiong, Wenjun
    Lagerstrom, Robert
    [J]. COMPUTERS & SECURITY, 2019, 84 : 53 - 69
  • [4] Bridge information modeling: A systematic literature review
    Monteiro, P. S.
    Wajdowicz, C. C.
    Santos, A. de P. L.
    Santos Filho, M. L.
    [J]. MAINTENANCE, MONITORING, SAFETY, RISK AND RESILIENCE OF BRIDGES AND BRIDGE NETWORKS, 2016, : 479 - 479
  • [5] Database Modeling for the Cloud: A Systematic Review of the Literature
    Garcia, Naomi
    Aguilar, Raul A.
    Diaz, Julio C.
    Flores, Brenda L.
    [J]. 2022 11TH INTERNATIONAL CONFERENCE ON SOFTWARE PROCESS IMPROVEMENT, CIMPS, 2022, : 85 - 94
  • [6] Reusability in goal modeling: A systematic literature review
    Duran, Mustafa Berk
    Mussbacher, Gunter
    [J]. INFORMATION AND SOFTWARE TECHNOLOGY, 2019, 110 : 156 - 173
  • [7] Sign Language Translation Systems: A Systematic Literature Review
    Boggaram, Ankith
    Boggaram, Aaptha
    Sharma, Aryan
    Ramanujan, Ashwin Srinivasa
    Bharathi, R.
    [J]. INTERNATIONAL JOURNAL OF SOFTWARE SCIENCE AND COMPUTATIONAL INTELLIGENCE-IJSSCI, 2022, 14 (01):
  • [8] Multimodality in the English language classroom: A systematic review of literature
    Lim, Fei Victor
    Toh, Weimin
    Nguyen, Thi Thu Ha
    [J]. LINGUISTICS AND EDUCATION, 2022, 69
  • [9] Language Disorders in Huntington Disease: A Systematic Literature Review
    Gagnon, Maude
    Barrette, Jasmine
    Macoir, Joel
    [J]. COGNITIVE AND BEHAVIORAL NEUROLOGY, 2018, 31 (04) : 179 - 192
  • [10] Systematic literature review of sentiment analysis in the Spanish language
    Osorio Angel, Sonia
    Pena Perez Negron, Adriana
    Espinoza-Valdez, Aurora
    [J]. DATA TECHNOLOGIES AND APPLICATIONS, 2021, 55 (04) : 461 - 479