Assessing the Strengths and Weaknesses of Large Language Models

被引:11
|
作者
Lappin, Shalom [1 ,2 ,3 ]
机构
[1] Queen Mary Univ London, Sch Elect Engn & Compuer Sci, London, England
[2] Univ Gothenburg, Ctr Linguist Theory & Studies Probabil, Gothenburg, Sweden
[3] Kings Coll London, Dept Informat, London, England
基金
瑞典研究理事会;
关键词
Deep learning; Transformers; Artifical intelligence; Natural language processing;
D O I
10.1007/s10849-023-09409-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). These are currently the focus of a lively discussion in both the scientific literature and the popular media. This discussion ranges from hyperbolic claims that attribute general intelligence and sentience to LLMs, to the skeptical view that these devices are no more than "stochastic parrots". I present an overview of some of the weak arguments that have been presented against LLMs, and I consider several of the more compelling criticisms of these devices. The former significantly underestimate the capacity of transformers to achieve subtle inductive inferences required for high levels of performance on complex, cognitively significant tasks. In some instances, these arguments misconstrue the nature of deep learning. The latter criticisms identify significant limitations in the way in which transformers learn and represent patterns in data. They also point out important differences between the procedures through which deep neural networks and humans acquire knowledge of natural language. It is necessary to look carefully at both sets of arguments in order to achieve a balanced assessment of the potential and the limitations of LLMs.
引用
收藏
页码:9 / 20
页数:12
相关论文
共 50 条
  • [31] Querying the Depths: Unveiling the Strengths andStruggles of Large Language Models in SPARQLGeneration
    Ghajari, Adrian
    Ros, Salvador
    Perez, Alvaro
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2024, (73): : 271 - 281
  • [32] In-Context Impersonation Reveals Large Language Models' Strengths and Biases
    Salewski, Leonard
    Alaniz, Stephan
    Rio-Torto, Isabel
    Schulz, Eric
    Akata, Zeynep
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [33] Play to Your Strengths: Collaborative Intelligence of Conventional Recommender Models and Large Language Models
    Xi, Yunjia
    Liu, Weiwen
    Lin, Jianghao
    Wu, Chuhan
    Chen, Bo
    Tang, Ruiming
    Zhang, Weinan
    Yu, Yong
    INFORMATION RETRIEVAL, CCIR 2024, 2025, 15418 : 1 - 13
  • [34] Assessing the strengths and weaknesses of voluntary programs to facilitate salmon restoration.
    Huffaker, RG
    JOURNAL OF AGRICULTURAL AND RESOURCE ECONOMICS, 2003, 28 (03): : 651 - 651
  • [35] Assessing "Global Lithuania": the strengths and weaknesses of Lithuanian diaspora engagement strategy
    Gudelis, Dangis
    Klimaviciute, Luka
    JOURNAL OF BALTIC STUDIES, 2016, 47 (03) : 325 - 348
  • [36] Understanding Learner Strengths and Weaknesses: Assessing Performance on an Integrated Writing Task
    Sawaki, Yasuyo
    Quinlan, Thomas
    Lee, Yong-Won
    LANGUAGE ASSESSMENT QUARTERLY, 2013, 10 (01) : 73 - 95
  • [37] MARKET MODELS FOR WATER-POLLUTION CONTROL - THEIR STRENGTHS AND WEAKNESSES
    ROSEACKERMAN, S
    PUBLIC POLICY, 1977, 25 (03): : 383 - 406
  • [38] Symbolic Integration by Integrating Learning Models With Different Strengths and Weaknesses
    Kubota, Hazumi
    Tokuoka, Yuta
    Yamada, Takahiro G.
    Funahashi, Akira
    IEEE ACCESS, 2022, 10 : 47000 - 47010
  • [39] Pediatric First Aid in Europe Strengths and Weaknesses of different Models
    Ehrich, J. H. H.
    PADIATRIE UND PADOLOGIE, 2013, 48 : 11 - 16
  • [40] Some strengths and weaknesses of overstress based elastic viscoplastic models
    Freitas, T. M. Bodas
    Potts, D.M.
    Zdravkovic, L.
    Springer Series in Geomechanics and Geoengineering, 2013, : 107 - 114