CrashTalk: Automated Generation of Precise, Human Readable, Descriptions of Software Security Bugs

被引:0
|
作者
James, Kedrian [1 ]
Valakuzhy, Kevin [2 ]
Snow, Kevin [3 ]
Monrose, Fabian [2 ]
机构
[1] Univ N Carolina, Chapel Hill, NC 27515 USA
[2] Georgia Inst Technol, Atlanta, GA 30332 USA
[3] Zeropoint Dynam, Chapel Hill, NC USA
关键词
Bug Localization; Debugging; Vulnerabilities;
D O I
10.1145/3626232.3653256
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Understanding the cause, consequences, and severity of a security bug are critical facets of the overall bug triaging and remediation process. Unfortunately, diagnosing failures is often a laborious process that requires developers to expend significant time and effort. While solutions have been proposed to help expedite the process of pinpointing the cause of a security bug, few proposals provide an explanation along with a diagnosis to make the bug discovery and triaging process less taxing. Moreover, even in cases where descriptions are provided, they are not guided by classification models that support precise descriptions of the flaw. We present an approach that uses static and dynamic analysis techniques to automatically infer the cause and consequences of a software crash and present diagnostic information following NIST's recently released Bugs Framework taxonomy. Specifically, starting from a crash, we generate a detailed and accessible English description of the failure along with its weakness types and severity, thereby easing the burden on developers and security analysts alike. To evaluate the effectiveness of our approach, we compare our ability to find fault locations and generate explanations compared to that of professional software developers by using a benchmark specifically designed to assist with realistic evaluation of tools in software engineering. In addition, using 33 real-world vulnerabilities we collected, we show that our approach correctly diagnoses over 94% of the failures and, in some cases, generates weakness types that are more specific than those that were originally assigned by the submitter or National Vulnerability Database analysts. We also generate initial vulnerability scores that can be used by project managers to assist with prioritizing bug fixes. On average, the overall process takes just over a minute, which is orders of magnitude faster than what professional developers can do.
引用
收藏
页码:337 / 347
页数:11
相关论文
共 6 条
  • [1] Automated Human-Readable Label Generation in Open Intent Discovery
    Anderson, Grant
    Hart, Emma
    Gkatzia, Dimitra
    Beaver, Ian
    INTERSPEECH 2024, 2024, : 3540 - 3544
  • [2] Automated Generation of Security-Centric Descriptions for Smart Contract Bytecode
    Pan, Yu
    Xu, Zhichao
    Li, Levi Taiji
    Yang, Yunhe
    Zhang, Mu
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 1244 - 1256
  • [3] Automated Generation of Human-readable Natural Arabic Text from RDF Data
    Touma, Roudy
    Hajj, Hazem
    El-Hajj, Wassim
    Shaban, Khaled
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (04)
  • [4] Machine learning technique for generation of human readable rules to detect software code smells in open-source software
    Stuti Tandon
    Vijay Kumar
    V. B. Singh
    Life Cycle Reliability and Safety Engineering, 2024, 13 (3) : 277 - 291
  • [5] Integrating Threat Modeling and Automated Test Case Generation into Industrialized Software Security Testing
    Marksteiner, Stefan
    Ramler, Rudolf
    Sochor, Hannes
    THIRD CENTRAL EUROPEAN CYBERSECURITY CONFERENCE (CECC 2019), 2019,
  • [6] The Generation of Software Security Scoring Systems Leveraging Human Expert Opinion
    Mell, Peter
    2022 IEEE 29TH ANNUAL SOFTWARE TECHNOLOGY CONFERENCE (STC 2022), 2022, : 116 - 124