Language Representation Models: An Overview

被引:10
|
作者
Schomacker, Thorben [1 ]
Tropmann-Frick, Marina [1 ]
机构
[1] Hamburg Univ Appl Sci, Dept Comp Sci, D-20099 Hamburg, Germany
关键词
natural language processing; neural networks; transformer; embeddings; multi-task learning; attention-based models; deep learning;
D O I
10.3390/e23111422
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
In the last few decades, text mining has been used to extract knowledge from free texts. Applying neural networks and deep learning to natural language processing (NLP) tasks has led to many accomplishments for real-world language problems over the years. The developments of the last five years have resulted in techniques that have allowed for the practical application of transfer learning in NLP. The advances in the field have been substantial, and the milestone of outperforming human baseline performance based on the general language understanding evaluation has been achieved. This paper implements a targeted literature review to outline, describe, explain, and put into context the crucial techniques that helped achieve this milestone. The research presented here is a targeted review of neural language models that present vital steps towards a general language representation model.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Towards Cross Language Process Model Reuse-A Language Independent Representation of Process Models
    Shahzad, Khurram
    Elias, Mturi
    Johannesson, Paul
    PRACTICE OF ENTERPRISE MODELING, PROCEEDINGS, 2009, 39 : 176 - 190
  • [22] An overview of mixed-effects statistical models for second language researchers
    Cunnings, Ian
    SECOND LANGUAGE RESEARCH, 2012, 28 (03) : 369 - 382
  • [23] CardioBERTpt: Transformer-based Models for Cardiology Language Representation in Portuguese
    Rubel Schneider, Elisa Terumi
    Gumiel, Yohan Bonescki
    Andrioli de Souza, Joao Vitor
    Mukai, Lilian Mie
    Silva e Oliveira, Lucas Emanuel
    Rebelo, Marina de Sa
    Gutierrez, Marco Antonio
    Krieger, Jose Eduardo
    Teodoro, Douglas
    Moro, Claudia
    Paraiso, Emerson Cabrera
    2023 IEEE 36TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS, CBMS, 2023, : 378 - 381
  • [24] Representation of Constituents in Neural Language Models: Coordination Phrase as a Case Study
    An, Aixiu
    Qian, Peng
    Wilcox, Ethan
    Levy, Roger
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 2888 - 2899
  • [25] REPRESENTATION AND REASONING MODELS FOR C3 ARCHITECTURE DESCRIPTION LANGUAGE
    Amirat, Abdelkrim
    Oussalah, Mourad
    ICEIS 2008: PROCEEDINGS OF THE TENTH INTERNATIONAL CONFERENCE ON ENTERPRISE INFORMATION SYSTEMS, VOL ISAS-2: INFORMATION SYSTEMS ANALYSIS AND SPECIFICATION, VOL 2, 2008, : 207 - +
  • [26] An Overview on the Existing Language Models for Prediction Systems as Writing Assistant Tools
    Ghayoomi, Masood
    Momtazi, Saeedeh
    2009 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC 2009), VOLS 1-9, 2009, : 5083 - 5087
  • [27] On the Generalization Abilities of Fine-Tuned Commonsense Language Representation Models
    Shen, Ke
    Kejriwal, Mayank
    ARTIFICIAL INTELLIGENCE XXXVIII, 2021, 13101 : 3 - 16
  • [28] Learning protein language contrastive models with multi-knowledge representation
    Xu, Wenjun
    Xia, Yingchun
    Sun, Bifan
    Zhao, Zihao
    Tang, Lianggui
    Zhou, Obo
    Wang, Qingyong
    Gu, Lichuan
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2025, 164
  • [29] Language-Models-as-a-Service: Overview of a New Paradigm and its Challenges
    La Malfa, Emanuele
    Petrov, Aleksandar
    Frieder, Simon
    Weinhuber, Christoph
    Burnell, Ryan
    Nazar, Raza
    Cohn, Anthony G.
    Shadbolt, Nigel
    Wooldridge, Michael
    Journal of Artificial Intelligence Research, 2024, 80 : 1497 - 1523
  • [30] Hierarchical Neural Language Models for Joint Representation of Streaming Documents and their Content
    Djuric, Nemanja
    Wu, Hao
    Radosavljevic, Vladan
    Grbovic, Mihajlo
    Bhamidipati, Narayan
    PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW 2015), 2015, : 248 - 255