Large Language Models: The Next Frontier for Variable Discovery within Metamorphic Testing

被引:2
|
作者
Tsigkanos, Christos [1 ]
Rani, Pooja [2 ]
Mueller, Sebastian [3 ]
Kehrer, Timo [1 ]
机构
[1] Univ Bern, Bern, Switzerland
[2] Univ Zurich, Zurich, Switzerland
[3] Humboldt Univ, Berlin, Germany
关键词
Metamorphic Testing; Large Language Models; Natural Language Processing; Scientific Software;
D O I
10.1109/SANER56733.2023.00070
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Metamorphic testing involves reasoning on necessary properties that a program under test should exhibit regarding multiple input and output variables. A general approach consists of extracting metamorphic relations from auxiliary artifacts such as user manuals or documentation, a strategy particularly fitting to testing scientific software. However, such software typically has large input-output spaces, and the fundamental prerequisite extracting variables of interest is an arduous and non-scalable process when performed manually. To this end, we devise a workflow around an autoregressive transformerbased Large Language Model (LLM) towards the extraction of variables from user manuals of scientific software. Our end-toend approach, besides a prompt specification consisting of fewshot examples by a human user, is fully automated, in contrast to current practice requiring human intervention. We showcase our LLM workflow over a real case, and compare variables extracted to ground truth manually labelled by experts. Our preliminary results show that our LLM-based workflow achieves an accuracy of 0.87, while successfully deriving 61.8% of variables as partial matches and 34.7% as exact matches.
引用
下载
收藏
页码:678 / 682
页数:5
相关论文
共 50 条
  • [21] LLMEffiChecker: Understanding and Testing Efficiency Degradation of Large Language Models
    Feng, Xiaoning
    Han, Xiaohong
    Chen, Simin
    Yang, Wei
    ACM Transactions on Software Engineering and Methodology, 2024, 33 (07)
  • [22] Leveraging Large Language Models to Improve REST API Testing
    Kim, Myeongsoo
    Stennett, Tyler
    Shah, Dhruv
    Sinha, Saurabh
    Orso, Alessandro
    2024 IEEE/ACM 46TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: NEW IDEAS AND EMERGING RESULTS, ICSE-NIER 2024, 2024, : 37 - 41
  • [23] Large Language Models for Code: Security Hardening and Adversarial Testing
    He, Jingxuan
    Vechev, Martin
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 1865 - 1879
  • [24] Exploring the new frontier of information extraction through large language models in urban analytics
    Crooks, Andrew
    Chen, Qingqing
    ENVIRONMENT AND PLANNING B-URBAN ANALYTICS AND CITY SCIENCE, 2024, 51 (03) : 565 - 569
  • [25] Large Language Models Meet Next-Generation Networking Technologies: A Review
    Hang, Ching-Nam
    Yu, Pei-Duo
    Morabito, Roberto
    Tan, Chee-Wei
    Future Internet, 2024, 16 (10):
  • [26] SheetCopilot: Bringing Software Productivity to the Next Level through Large Language Models
    Li, Hongxin
    Su, Jingran
    Chen, Yuntao
    Li, Qing
    Zhang, Zhaoxiang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [27] Emergent effects of scaling on the functional hierarchies within large language models
    Bogdan, Paul C.
    arXiv,
  • [28] Correctable Landmark Discovery Via Large Models for Vision-Language Navigation
    Lin B.
    Nie Y.
    Wei Z.
    Zhu Y.
    Xu H.
    Ma S.
    Liu J.
    Liang X.
    IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46 (12) : 1 - 14
  • [29] Artificial intelligence enabled ChatGPT and large language models in drug target discovery, drug discovery, and development
    Chakraborty, Chiranjib
    Bhattacharya, Manojit
    Lee, Sang-Soo
    MOLECULAR THERAPY-NUCLEIC ACIDS, 2023, 33 : 866 - 868
  • [30] Disambiguate Entity Matching using Large Language Models through Relation Discovery
    Huang, Zezhou
    1st Workshop on Governance, Understanding and Integration of Data for Effective and Responsible AI, GUIDE-AI 2024, Co-located with SIGMOD 2024, : 36 - 39