Detecting Data Races in OpenMP with Deep Learning and Large Language Models

被引:0
|
作者
Alsofyani, May [1 ]
Wang, Liqiang [1 ]
机构
[1] Univ Cent Florida, Dept Comp Sci, Orlando, FL 32816 USA
关键词
data race; race condition; bug detection; OpenMP; transformer encoder; large language model; CodeBERTa; GPT-4; Turbo;
D O I
10.1145/3677333.3678160
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based neural network models are increasingly employed to handle software engineering issues, such as bug localization and program repair. These models, equipped with a self-attention mechanism, excel at understanding source code context and semantics. Recently, large language models (LLMs) have emerged as a promising alternative for analyzing and understanding code structure. In this paper, we propose two novel methods for detecting data race bugs in OpenMP programs. The first method is based on a transformer encoder trained from scratch. The second method leverages LLMs, specifically extending GPT-4 Turbo through the use of prompt engineering and fine-tuning techniques. For training and testing our approach, we utilized two datasets comprising different OpenMP directives. Our experiments show that the transformer encoder achieves competitive accuracy compared to LLMs, whether through fine-tuning or prompt engineering techniques. This performance may be attributed to the complexity of many OpenMP directives and the limited availability of labeled datasets.
引用
收藏
页码:96 / 103
页数:8
相关论文
共 50 条
  • [31] Detecting the Speaker Language Using CNN Deep Learning Algorithm
    Rammo F.M.
    Al-Hamdani M.N.
    Iraqi Journal for Computer Science and Mathematics, 2022, 3 (01): : 43 - 52
  • [32] Detecting racism and xenophobia using deep learning models on Twitter data: CNN, LSTM and BERT
    Alberto Benitez-Andrades, Jose
    Gonzalez-Jimenez, Alvaro
    Lopez-Brea, Alvaro
    Aveleira-Mata, Jose
    Alija-Perez, Jose-Manuel
    Teresa Garcia-Ordas, Maria
    PEERJ COMPUTER SCIENCE, 2022, 8
  • [33] Detecting racism and xenophobia using deep learning models on Twitter data: CNN, LSTM and BERT
    Benítez-Andrades J.A.
    González-Jiménez Á.
    López-Brea Á.
    Aveleira-Mata J.
    Alija-Pérez J.-M.
    García-Ordás M.T.
    PeerJ Computer Science, 2022, 8
  • [34] Tool learning with large language models: a survey
    Changle Qu
    Sunhao Dai
    Xiaochi Wei
    Hengyi Cai
    Shuaiqiang Wang
    Dawei Yin
    Jun Xu
    Ji-rong Wen
    Frontiers of Computer Science, 2025, 19 (8)
  • [35] Shortcut Learning of Large Language Models in Natural Language Understanding
    Du, Mengnan
    He, Fengxiang
    Zou, Na
    Tao, Dacheng
    Hu, Xia
    COMMUNICATIONS OF THE ACM, 2024, 67 (01) : 110 - 120
  • [36] Large Language Models Demonstrate the Potential of Statistical Learning in Language
    Contreras Kallens, Pablo
    Kristensen-McLachlan, Ross Deans
    Christiansen, Morten H.
    COGNITIVE SCIENCE, 2023, 47 (03) : e13256
  • [37] An Investigation of Applying Large Language Models to Spoken Language Learning
    Gao, Yingming
    Nuchged, Baorian
    Li, Ya
    Peng, Linkai
    APPLIED SCIENCES-BASEL, 2024, 14 (01):
  • [38] Federated deep learning models for detecting RPL attacks on large-scale hybrid IoT networks
    Albishari, Mohammed
    Li, Mingchu
    Ayoubi, Majid
    Alsanabani, Ala
    Tian, Jiyu
    Computer Networks, 2024, 254
  • [39] FRACTAL-INSPIRED SENTIMENT ANALYSIS: EVALUATION OF LARGE LANGUAGE MODELS AND DEEP LEARNING METHODS
    Alsagri, Hatoon S.
    Sohail, Shahab Saquib
    FRACTALS-COMPLEX GEOMETRY PATTERNS AND SCALING IN NATURE AND SOCIETY, 2024,