Language Representation Models: An Overview

被引:10
|
作者
Schomacker, Thorben [1 ]
Tropmann-Frick, Marina [1 ]
机构
[1] Hamburg Univ Appl Sci, Dept Comp Sci, D-20099 Hamburg, Germany
关键词
natural language processing; neural networks; transformer; embeddings; multi-task learning; attention-based models; deep learning;
D O I
10.3390/e23111422
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
In the last few decades, text mining has been used to extract knowledge from free texts. Applying neural networks and deep learning to natural language processing (NLP) tasks has led to many accomplishments for real-world language problems over the years. The developments of the last five years have resulted in techniques that have allowed for the practical application of transfer learning in NLP. The advances in the field have been substantial, and the milestone of outperforming human baseline performance based on the general language understanding evaluation has been achieved. This paper implements a targeted literature review to outline, describe, explain, and put into context the crucial techniques that helped achieve this milestone. The research presented here is a targeted review of neural language models that present vital steps towards a general language representation model.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Graphical Models for Preference Representation: An Overview
    Ben Amor, Nahla
    Dubois, Didier
    Gouider, Hela
    Prade, Henri
    SCALABLE UNCERTAINTY MANAGEMENT, SUM 2016, 2016, 9858 : 96 - 111
  • [2] Representation of the language and models of language change: grammaticalization as perspective
    Feltgen, Quentin
    Fagard, Benjamin
    Nadal, Jean-Pierre
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2014, 55 (03): : 47 - 71
  • [3] Can language representation models think in bets?
    Tang, Zhisheng
    Kejriwal, Mayank
    ROYAL SOCIETY OPEN SCIENCE, 2023, 10 (03):
  • [4] An Overview of Language Models: Recent Developments and Outlook
    Wei, Chengwei
    Wang, Yun-Cheng
    Wang, Bin
    Kuo, C. -C. Jay
    APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, 2024, 13 (02)
  • [5] Monotonic Representation of Numeric Properties in Language Models
    Heinzerling, Benjamin
    Inui, Kentaro
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 175 - 195
  • [6] Overestimation of Syntactic Representation in Neural Language Models
    Kodner, Jordan
    Gupta, Nitish
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 1757 - 1762
  • [7] An overview of diagnostics and therapeutics using large language models
    Malgaroli, Matteo
    Mcduff, Daniel
    JOURNAL OF TRAUMATIC STRESS, 2024, 37 (05) : 754 - 760
  • [8] The Less the Merrier? Investigating Language Representation in Multilingual Models
    Nigatu, Hellina Hailu
    Tonja, Atnafu Lambebo
    Kalita, Jugal
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12572 - 12589
  • [9] A Comprehensive Survey on Word Representation Models: From Classical to State-of-the-Art Word Representation Language Models
    Naseem, Usman
    Razzak, Imran
    Khan, Shah Khalid
    Prasad, Mukesh
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (05)
  • [10] Natural Language Processing: An Overview of Models, Transformers and Applied Practices
    Canchila, Santiago
    Meneses-Eraso, Carlos
    Casanoves-Boix, Javier
    Cortes-Pellicer, Pascual
    Castello-Sirvent, Fernando
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2024, 21 (03)