Towards Safer Large Language Models (LLMs)

被引:0
|
作者
Lawrence, Carolin [1 ]
Bifulco, Roberto [1 ]
Gashteovski, Kiril [1 ]
Hung, Chia-Chien [1 ]
Ben Rim, Wiem [1 ]
Shaker, Ammar [1 ]
Oyamada, Masafumi [2 ]
Sadamasa, Kunihiko [2 ]
Enomoto, Masafumi [2 ]
Takeoka, Kunihiro [2 ]
机构
[1] NEC Laboratories Europe, Germany
[2] Data Science Laboratories
来源
NEC Technical Journal | 2024年 / 17卷 / 02期
关键词
Computational linguistics - Risk assessment;
D O I
暂无
中图分类号
学科分类号
摘要
Large Language Models (LLMs) are revolutionizing our world. They have impressive textual capabilities that will fundamentally change how human users can interact with intelligent systems. Nonetheless, they also still have a series of limitations that are important to keep in mind when working with LLMs. We explore how these limitations can be addressed from two different angles. First, we look at options that are currently already available, which include (1) assessing the risk of a use case, (2) prompting a LLM to deliver explanations and (3) encasing LLMs in a human-centred system design. Second, we look at technologies that we are currently developing, which will be able to (1) more accurately assess the quality of an LLM for a high-risk domain, (2) explain the generated LLM output by linking to the input and (3) fact check the generated LLM output against external trustworthy sources. © 2024 NEC Mediaproducts. All rights reserved.
引用
下载
收藏
页码:64 / 74
相关论文
共 50 条
  • [31] Harnessing the Power of Large Language Models (LLMs) for Electronic Health Records (EHRs) Optimization
    Nashwan, Abdulqadir J.
    Abujaber, Ahmad A.
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (07)
  • [32] Large Language Models (LLMs) in Engineering Education: A Systematic Review and Suggestions for Practical Adoption
    Filippi, Stefano
    Motyl, Barbara
    INFORMATION, 2024, 15 (06)
  • [33] LEVERAGING LARGE LANGUAGE MODELS (LLMS) FOR CLASSIFYING PEER REVIEWED PUBLICATIONS FOR LITERATURE REVIEW
    Lee, S. H.
    Chacko, A.
    Yankovsky, A.
    VALUE IN HEALTH, 2024, 27 (06) : S262 - S262
  • [34] Can LLMs Master Math? Investigating Large Language Models on Math Stack Exchange
    Satpute, Ankit
    Giessing, Noah
    Greiner-Petter, Andre
    Schubotz, Moritz
    Teschke, Olaf
    Aizawa, Akiko
    Gipp, Bela
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2316 - 2320
  • [35] A survey on augmenting knowledge graphs (KGs) with large language models (LLMs): models, evaluation metrics, benchmarks, and challenges
    Ibrahim, Nourhan
    Aboulela, Samar
    Ibrahim, Ahmed
    Kashef, Rasha
    Discover Artificial Intelligence, 2024, 4 (01):
  • [36] GPT, large language models (LLMs) and generative artificial intelligence (GAI) models in geospatial science: a systematic review
    Wang, Siqin
    Hu, Tao
    Xiao, Huang
    Li, Yun
    Zhang, Ce
    Ning, Huan
    Zhu, Rui
    Li, Zhenlong
    Ye, Xinyue
    INTERNATIONAL JOURNAL OF DIGITAL EARTH, 2024, 17 (01)
  • [37] Generating Specifications from Requirements Documents for Smart Devices Using Large Language Models (LLMs)
    Lutze, Rainer
    Waldhoer, Klemens
    HUMAN-COMPUTER INTERACTION, PT I, HCI 2024, 2024, 14684 : 94 - 108
  • [38] The Role of Large Language Models (LLMs) in Providing Triage for Maxillofacial Trauma Cases: A Preliminary Study
    Frosolini, Andrea
    Catarzi, Lisa
    Benedetti, Simone
    Latini, Linda
    Chisci, Glauco
    Franz, Leonardo
    Gennaro, Paolo
    Gabriele, Guido
    DIAGNOSTICS, 2024, 14 (08)
  • [39] Beyond Probabilities: Unveiling the Delicate Dance of Large Language Models (LLMs) and AI-Hallucination
    Hamid, Oussama H.
    2024 IEEE CONFERENCE ON COGNITIVE AND COMPUTATIONAL ASPECTS OF SITUATION MANAGEMENT, COGSIMA, 2024, : 85 - 90
  • [40] PromSec: Prompt Optimization for Secure Generation of Functional Source Code with Large Language Models (LLMs)
    Nazzal, Mahmoud
    Khalil, Issa
    Khreishah, Abdallah
    Phan, NhatHai
    CCS 2024 - Proceedings of the 2024 ACM SIGSAC Conference on Computer and Communications Security, : 2266 - 2279