Evaluating the Efficacy of Large Language Models in Identifying Phishing Attempts

被引:0
|
作者
Patel, Het [1 ]
Reiman, Umair [1 ]
Iqbal, Farkhund [2 ]
机构
[1] Western Univ, Dept Comp Sci, London, ON, Canada
[2] Zayed Univ, Coll Technol Innovat, Dubai, U Arab Emirates
关键词
Phishing Email Detection; Large Language Models (LLMs); General Pretrained Transformer (GPT); Bidirectional Encoder Representations from Transformers (BERT); Natural Processing Language (NPL); Social Engineering;
D O I
10.1109/HSI61632.2024.10613528
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Phishing, a prevalent cybercrime tactic for decades, remains a significant threat in today's digital world. By leveraging clever social engineering elements and modern technology, cybercrime targets many individuals, businesses, and organizations to exploit trust and security. These cyberattackers are often disguised in many trustworthy forms to appear as legitimate sources. By cleverly using psychological elements like urgency, fear, social proof, and other manipulative strategies, phishers can lure individuals into revealing sensitive and personalized information. Building on this pervasive issue within modern technology, this paper will aim to analyze the effectiveness of 15 Large Language Models (LLMs) in detecting phishing attempts, specifically focusing on a randomized set of "419 Scam" emails. The objective is to determine which LLMs can accurately detect phishing emails by analyzing a text file containing email metadata based on predefined criteria. The experiment concluded that the following models, ChatGPT 3.5, GPT-3.5-Turbo-Instruct, and ChatGPT, were the most effective in detecting phishing emails.
引用
收藏
页数:7
相关论文
共 50 条
  • [1] Devising and Detecting Phishing Emails Using Large Language Models
    Heiding, Fredrik
    Schneier, Bruce
    Vishwanath, Arun
    Bernstein, Jeremy
    Park, Peter S.
    [J]. IEEE ACCESS, 2024, 12 : 42131 - 42146
  • [2] ChatPhishDetector: Detecting Phishing Sites Using Large Language Models
    Koide, Takashi
    Nakano, Hiroki
    Chiba, Daiki
    [J]. IEEE Access, 2024, 12 : 154381 - 154400
  • [3] Evaluating large language models for annotating proteins
    Vitale, Rosario
    Bugnon, Leandro A.
    Fenoy, Emilio Luis
    Milone, Diego H.
    Stegmayer, Georgina
    [J]. BRIEFINGS IN BIOINFORMATICS, 2024, 25 (03)
  • [4] A bilingual benchmark for evaluating large language models
    Alkaoud, Mohamed
    [J]. PEERJ COMPUTER SCIENCE, 2024, 10
  • [5] Evaluating large language models as agents in the clinic
    Nikita Mehandru
    Brenda Y. Miao
    Eduardo Rodriguez Almaraz
    Madhumita Sushil
    Atul J. Butte
    Ahmed Alaa
    [J]. npj Digital Medicine, 7
  • [6] Evaluating large language models as agents in the clinic
    Mehandru, Nikita
    Miao, Brenda Y.
    Almaraz, Eduardo Rodriguez
    Sushil, Madhumita
    Butte, Atul J.
    Alaa, Ahmed
    [J]. NPJ DIGITAL MEDICINE, 2024, 7 (01)
  • [7] Evaluating Intelligence and Knowledge in Large Language Models
    Bianchini, Francesco
    [J]. TOPOI-AN INTERNATIONAL REVIEW OF PHILOSOPHY, 2024,
  • [8] From Chatbots to Phishbots?: Phishing Scam Generation in Commercial Large Language Models
    Roy, Sayak Saha
    Thota, Poojitha
    Naragam, Krishna Vamsi
    Nilizadeh, Shirin
    [J]. Proceedings - IEEE Symposium on Security and Privacy, 2024, : 36 - 54
  • [9] Evaluating large language models on medical evidence summarization
    Liyan Tang
    Zhaoyi Sun
    Betina Idnay
    Jordan G. Nestor
    Ali Soroush
    Pierre A. Elias
    Ziyang Xu
    Ying Ding
    Greg Durrett
    Justin F. Rousseau
    Chunhua Weng
    Yifan Peng
    [J]. npj Digital Medicine, 6
  • [10] Evaluating large language models on medical evidence summarization
    Tang, Liyan
    Sun, Zhaoyi
    Idnay, Betina
    Nestor, Jordan G.
    Soroush, Ali
    Elias, Pierre A.
    Xu, Ziyang
    Ding, Ying
    Durrett, Greg
    Rousseau, Justin F.
    Weng, Chunhua
    Peng, Yifan
    [J]. NPJ DIGITAL MEDICINE, 2023, 6 (01)