A Review on Large Language Models: Architectures, Applications, Taxonomies, Open Issues and Challenges

被引:19
|
作者
Raiaan, Mohaimenul Azam Khan [1 ]
Mukta, Md. Saddam Hossain [2 ]
Fatema, Kaniz [3 ]
Fahad, Nur Mohammad [1 ]
Sakib, Sadman [1 ]
Mim, Most Marufatul Jannat [1 ]
Ahmad, Jubaer [1 ]
Ali, Mohammed Eunus [4 ]
Azam, Sami [3 ]
机构
[1] United Int Univ, Dept Comp Sci & Engn, Dhaka 1212, Bangladesh
[2] Lappeenranta Lahti Univ Technol, LUT Sch Engn Sci, Lappeenranta 53850, Finland
[3] Charles Darwin Univ, Fac Sci & Technol, Casuarina, NT 0909, Australia
[4] Bangladesh Univ Engn & Technol BUET, Dept CSE, Dhaka 1000, Bangladesh
关键词
Cognition; Artificial intelligence; Transformers; Training; Taxonomy; Task analysis; Surveys; Natural language processing; Question answering (information retrieval); Information analysis; Linguistics; Large language models (LLM); natural language processing (NLP); artificial intelligence; transformer; pre-trained models; taxonomy; application; GPT-4; BIAS;
D O I
10.1109/ACCESS.2024.3365742
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large Language Models (LLMs) recently demonstrated extraordinary capability in various natural language processing (NLP) tasks including language translation, text generation, question answering, etc. Moreover, LLMs are new and essential part of computerized language processing, having the ability to understand complex verbal patterns and generate coherent and appropriate replies in a given context. Though this success of LLMs has prompted a substantial increase in research contributions, rapid growth has made it difficult to understand the overall impact of these improvements. Since a plethora of research on LLMs have been appeared within a short time, it is quite impossible to track all of these and get an overview of the current state of research in this area. Consequently, the research community would benefit from a short but thorough review of the recent changes in this area. This article thoroughly overviews LLMs, including their history, architectures, transformers, resources, training methods, applications, impacts, challenges, etc. This paper begins by discussing the fundamental concepts of LLMs with its traditional pipeline of the LLMs training phase. Then the paper provides an overview of the existing works, the history of LLMs, their evolution over time, the architecture of transformers in LLMs, the different resources of LLMs, and the different training methods that have been used to train them. The paper also demonstrates the datasets utilized in the studies. After that, the paper discusses the wide range of applications of LLMs, including biomedical and healthcare, education, social, business, and agriculture. The study also illustrates how LLMs create an impact on society and shape the future of AI and how they can be used to solve real-world problems. Finally, the paper also explores open issues and challenges to deploy LLMs in real-world scenario. Our review paper aims to help practitioners, researchers, and experts thoroughly understand the evolution of LLMs, pre-trained architectures, applications, challenges, and future goals.
引用
收藏
页码:26839 / 26874
页数:36
相关论文
共 50 条
  • [1] A review on deep learning approaches in healthcare systems: Taxonomies, challenges, and open issues
    Shamshirband, Shahab
    Fathi, Mahdis
    Dehzangi, Abdollah
    Chronopoulos, Anthony Theodore
    Alinejad-Rokny, Hamid
    [J]. JOURNAL OF BIOMEDICAL INFORMATICS, 2021, 113 (113)
  • [2] Machine learning and internet of things applications in enterprise architectures: Solutions, challenges, and open issues
    Rehman, Zubaida
    Tariq, Noshina
    Moqurrab, Syed Atif
    Yoo, Joon
    Srivastava, Gautam
    [J]. EXPERT SYSTEMS, 2024, 41 (01)
  • [3] Embracing Large Language Models for Medical Applications: Opportunities and Challenges
    Karabacak, Mert
    Margetis, Konstantinos
    [J]. CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (05)
  • [4] Large language models in health care: Development, applications, and challenges
    Yang, Rui
    Tan, Ting Fang
    Lu, Wei
    Thirunavukarasu, Arun James
    Ting, Daniel Shu Wei
    Liu, Nan
    [J]. HEALTH CARE SCIENCE, 2023, 2 (04): : 255 - 263
  • [5] Applications of large language models in psychiatry: a systematic review
    Omar, Mahmud
    Soffer, Shelly
    Charney, Alexander W.
    Landi, Isotta
    Nadkarni, Girish N.
    Klang, Eyal
    [J]. FRONTIERS IN PSYCHIATRY, 2024, 15
  • [6] Large language models in medical and healthcare fields: applications, advances, and challenges
    Wang, Dandan
    Zhang, Shiqing
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (11)
  • [7] Clinical and Surgical Applications of Large Language Models: A Systematic Review
    Pressman, Sophia M.
    Borna, Sahar
    Gomez-Cabello, Cesar A.
    Haider, Syed Ali
    Haider, Clifton R.
    Forte, Antonio Jorge
    [J]. JOURNAL OF CLINICAL MEDICINE, 2024, 13 (11)
  • [8] Metaheuristic algorithms and their applications in wireless sensor networks: review, open issues, and challenges
    Houssein, Essam H.
    Saad, Mohammed R.
    Djenouri, Youcef
    Hu, Gang
    Ali, Abdelmgeid A.
    Shaban, Hassan
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2024, 27 (10): : 13643 - 13673
  • [9] A Review of Fog Computing and Machine Learning: Concepts, Applications, Challenges, and Open Issues
    Abdulkareem, Karrar Hameed
    Mohammed, Mazin Abed
    Gunasekaran, Saraswathy Shamini
    Al-Mhiqani, Mohammed Nasser
    Mutlag, Ammar Awad
    Mostafa, Salama A.
    Ali, Nabeel Salih
    Ibrahim, Dheyaa Ahmed
    [J]. IEEE ACCESS, 2019, 7 : 153123 - 153140
  • [10] A Review of Current Trends, Techniques, and Challenges in Large Language Models (LLMs)
    Patil, Rajvardhan
    Gudivada, Venkat
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (05):