Leveraging Large Language Models for Efficient Failure Analysis in Game Development

被引:0
|
作者
Marini, Leonardo [1 ]
Gisslen, Linus [2 ]
Sestini, Alessandro [2 ]
机构
[1] Frostbite, Stockholm, Sweden
[2] SEED Elect Arts EA, Redwood City, CA USA
关键词
Natural language processing; Validation; Tracing; Games; Software Quality; Software development;
D O I
10.1109/CoG60054.2024.10645540
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In games, and more generally in the field of software development, early detection of bugs is vital to maintain a high quality of the final product. Automated tests are a powerful tool that can catch a problem earlier in development by executing periodically. As an example, when new code is submitted to the code base, a new automated test verifies these changes. However, identifying the specific change responsible for a test failure becomes harder when dealing with batches of changes especially in the case of a large-scale project such as a AAA game, where thousands of people contribute to a single code base. This paper proposes a new approach to automatically identify which change in the code caused a test to fail. The method leverages Large Language Models (LLMs) to associate error messages with the corresponding code changes causing the failure. We investigate the effectiveness of our approach with quantitative and qualitative evaluations. Our approach reaches an accuracy of 71% in our newly created dataset, which comprises issues reported by developers at EA over a period of one year. We further evaluated our model through a user study to assess the utility and usability of the tool from a developer perspective, resulting in a significant reduction in time - up to 60% - spent investigating issues.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Leveraging Large Language Models for Enhanced Classification and Analysis: Fire Incidents Case Study
    Alkhammash, Eman H.
    FIRE-SWITZERLAND, 2025, 8 (01):
  • [32] Leveraging Large Language Models for QA Dialogue Dataset Construction and Analysis in Public Services
    Wu, Chaomin
    Wu, Di
    Pan, Yushan
    Wang, Hao
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT I, NLPCC 2024, 2025, 15359 : 56 - 68
  • [33] LEVERAGING LARGE LANGUAGE MODELS FOR EFFICIENT OBSERVATIONAL RESEARCH PROTOCOL DRAFTING: INSIGHTS AND BEST PRACTICES FOR OPENLLAMA
    Texeira, B. Y.
    Merrill, C.
    Gao, W.
    Gao, C.
    Bao, Y.
    Anstatt, D.
    VALUE IN HEALTH, 2024, 27 (06) : S276 - S276
  • [34] Leveraging large language models: transforming scholarly publishing for the better
    Fortier, Lisa A.
    AMERICAN JOURNAL OF VETERINARY RESEARCH, 2023, 84 (08) : 1 - 2
  • [35] Leveraging foundation and large language models in medical artificial intelligence
    Wong Io Nam
    Monteiro Olivia
    BaptistaHon Daniel T
    Wang Kai
    Lu Wenyang
    Sun Zhuo
    Nie Sheng
    Yin Yun
    中华医学杂志英文版, 2024, 137 (21)
  • [36] Leveraging large language models to monitor climate technology innovation
    Toetzke, Malte
    Probst, Benedict
    Feuerriegel, Stefan
    ENVIRONMENTAL RESEARCH LETTERS, 2023, 18 (09)
  • [37] Position Paper: Leveraging Large Language Models for Cybersecurity Compliance
    Salman, Ahmed
    Creese, Sadie
    Goldsmith, Michael
    9TH IEEE EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS, EUROS&PW 2024, 2024, : 496 - 503
  • [38] Leveraging foundation and large language models in medical artificial intelligence
    Wong, Io Nam
    Monteiro, Olivia
    Baptista-Hon, Daniel T.
    Wang, Kai
    Lu, Wenyang
    Sun, Zhuo
    Nie, Sheng
    Yin, Yun
    CHINESE MEDICAL JOURNAL, 2024, 137 (21) : 2529 - 2539
  • [39] Leveraging Large Language Models to Detect npm Malicious Packages
    Zahan, Nusrat
    Burckhardt, Philipp
    Lysenko, Mikola
    Aboukhadijeh, Feross
    Williams, Laurie
    arXiv,
  • [40] Leveraging Large Language Models for Activity Recognition in Smart Environments
    Cleland, Ian
    Nugent, Luke
    Cruciani, Federico
    Nugent, Chris
    2024 INTERNATIONAL CONFERENCE ON ACTIVITY AND BEHAVIOR COMPUTING, ABC 2024, 2024,