Assessing the Strengths and Weaknesses of Large Language Models

被引:11
|
作者
Lappin, Shalom [1 ,2 ,3 ]
机构
[1] Queen Mary Univ London, Sch Elect Engn & Compuer Sci, London, England
[2] Univ Gothenburg, Ctr Linguist Theory & Studies Probabil, Gothenburg, Sweden
[3] Kings Coll London, Dept Informat, London, England
基金
瑞典研究理事会;
关键词
Deep learning; Transformers; Artifical intelligence; Natural language processing;
D O I
10.1007/s10849-023-09409-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). These are currently the focus of a lively discussion in both the scientific literature and the popular media. This discussion ranges from hyperbolic claims that attribute general intelligence and sentience to LLMs, to the skeptical view that these devices are no more than "stochastic parrots". I present an overview of some of the weak arguments that have been presented against LLMs, and I consider several of the more compelling criticisms of these devices. The former significantly underestimate the capacity of transformers to achieve subtle inductive inferences required for high levels of performance on complex, cognitively significant tasks. In some instances, these arguments misconstrue the nature of deep learning. The latter criticisms identify significant limitations in the way in which transformers learn and represent patterns in data. They also point out important differences between the procedures through which deep neural networks and humans acquire knowledge of natural language. It is necessary to look carefully at both sets of arguments in order to achieve a balanced assessment of the potential and the limitations of LLMs.
引用
收藏
页码:9 / 20
页数:12
相关论文
共 50 条
  • [1] Assessing the Strengths and Weaknesses of Large Language Models
    Shalom Lappin
    Journal of Logic, Language and Information, 2024, 33 : 9 - 20
  • [2] Evaluating the strengths and weaknesses of large language models in answering neurophysiology questions
    Shojaee-Mend, Hassan
    Mohebbati, Reza
    Amiri, Mostafa
    Atarodi, Alireza
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [3] ChatGPT as an inventor: eliciting the strengths and weaknesses of current large language models against humans in engineering design
    Ege, Daniel N.
    Ovrebo, Henrik H.
    Stubberud, Vegar
    Berg, Martin F.
    Elverum, Christer
    Steinert, Martin
    Vestad, Havard
    AI EDAM-ARTIFICIAL INTELLIGENCE FOR ENGINEERING DESIGN ANALYSIS AND MANUFACTURING, 2025, 39
  • [4] Glioblastoma preclinical models: Strengths and weaknesses
    Pasupuleti, Vasavi
    Vora, Lalitkumar
    Prasad, Renuka
    Nandakumar, D. N.
    Khatri, Dharmendra Kumar
    BIOCHIMICA ET BIOPHYSICA ACTA-REVIEWS ON CANCER, 2024, 1879 (01):
  • [5] CORPORATE APPRAISAL - ASSESSING COMPANY STRENGTHS AND WEAKNESSES
    HUSSEY, DE
    LONG RANGE PLANNING, 1968, 1 (02) : 19 - 25
  • [6] Assessing national strengths and weaknesses in research fields
    Abramo, Giovanni
    D'Angelo, Ciriaco Andrea
    JOURNAL OF INFORMETRICS, 2014, 8 (03) : 766 - 775
  • [7] STRENGTHS AND WEAKNESSES OF ASSESSING VITAMIN CONTENT OF FOODS
    EITENMILLER, RR
    JOURNAL OF FOOD QUALITY, 1990, 13 (01) : 7 - 20
  • [8] Backward readers’ awareness of language: Strengths and weaknesses
    Peter Bryant
    Terezinha Nunes
    Miriam Bindman
    European Journal of Psychology of Education, 1997, 12 : 357 - 372
  • [9] Backward readers' awareness of language: Strengths and weaknesses
    Bryant, P
    Nunes, T
    Bindman, M
    EUROPEAN JOURNAL OF PSYCHOLOGY OF EDUCATION, 1997, 12 (04) : 357 - 372
  • [10] Animal Models for Influenza Research: Strengths and Weaknesses
    Nguyen, Thi-Quyen
    Rollon, Rare
    Choi, Young-Ki
    VIRUSES-BASEL, 2021, 13 (06):