Devising and Detecting Phishing Emails Using Large Language Models

被引:9
|
作者
Heiding, Fredrik [1 ,2 ]
Schneier, Bruce [3 ]
Vishwanath, Arun [4 ]
Bernstein, Jeremy [5 ]
Park, Peter S. [5 ]
机构
[1] Harvard Univ, Harvard John A Paulson Sch Engn & Appl Sci, Cambridge, MA 02138 USA
[2] KTH Royal Inst Technol, S-11428 Stockholm, Sweden
[3] Harvard Univ, Harvard Kennedy Sch, Cambridge, MA 02138 USA
[4] Avant Res Grp, Buffalo, NY 14214 USA
[5] MIT, Cambridge, MA 02139 USA
关键词
Phishing; large language models; social engineering; artificial intelligence;
D O I
10.1109/ACCESS.2024.3375882
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
AI programs, built using large language models, make it possible to automatically create phishing emails based on a few data points about a user. The V-Triad is a set of rules for manually designing phishing emails to exploit our cognitive heuristics and biases. In this study, we compare the performance of phishing emails created automatically by GPT-4 and manually using the V-Triad. We also combine GPT-4 with the V-Triad to assess their combined potential. A fourth group, exposed to generic phishing emails, was our control group. We use a red teaming approach by simulating attackers and emailing 112 participants recruited for the study. The control group emails received a click-through rate between 19-28%, the GPT-generated emails 30-44%, emails generated by the V-Triad 69-79%, and emails generated by GPT and the V-Triad 43-81%. Each participant was asked to explain why they pressed or did not press a link in the email. These answers often contradict each other, highlighting the importance of personal differences. Next, we used four popular large language models (GPT, Claude, PaLM, and LLaMA) to detect the intention of phishing emails and compare the results to human detection. The language models demonstrated a strong ability to detect malicious intent, even in non-obvious phishing emails. They sometimes surpassed human detection, although often being slightly less accurate than humans. Finally, we analyze of the economic aspects of AI-enabled phishing attacks, showing how large language models increase the incentives of phishing and spear phishing by reducing their costs.
引用
收藏
页码:42131 / 42146
页数:16
相关论文
共 50 条
  • [21] From Chatbots to Phishbots?: Phishing Scam Generation in Commercial Large Language Models
    Roy, Sayak Saha
    Thota, Poojitha
    Naragam, Krishna Vamsi
    Nilizadeh, Shirin
    45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP 2024, 2024, : 36 - 54
  • [22] “Alexa, What’s a Phishing Email?”: Training users to spot phishing emails using a voice assistant
    Filipo Sharevski
    Peter Jachim
    EURASIP Journal on Information Security, 2022
  • [23] Preventing and Detecting Misinformation Generated by Large Language Models
    Liu, Aiwei
    Sheng, Qiang
    Hu, Xuming
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 3001 - 3004
  • [24] Detecting and Preventing Hallucinations in Large Vision Language Models
    Gunjal, Anisha
    Yin, Jihan
    Bas, Erhan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18135 - 18143
  • [25] Detecting ham and spam emails using feature union and supervised machine learning models
    Rustam, Furqan
    Saher, Najia
    Mehmood, Arif
    Lee, Ernesto
    Washington, Sandrilla
    Ashraf, Imran
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (17) : 26545 - 26561
  • [26] Detecting ham and spam emails using feature union and supervised machine learning models
    Furqan Rustam
    Najia Saher
    Arif Mehmood
    Ernesto Lee
    Sandrilla Washington
    Imran Ashraf
    Multimedia Tools and Applications, 2023, 82 : 26545 - 26561
  • [27] Disguised Executable Files in Spear-Phishing Emails: Detecting the Point of Entry in Advanced Persistent Threat
    Ghafir, Ibrahim
    Prenosil, Vaclav
    Hammoudeh, Mohammad
    Aparicio-Navarro, Francisco J.
    Rabie, Khaled
    Jabban, Ahmad
    ICFNDS'18: PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON FUTURE NETWORKS AND DISTRIBUTED SYSTEMS, 2018,
  • [28] Detecting AI-Generated Code Assignments Using Perplexity of Large Language Models
    Xu, Zhenyu
    Sheng, Victor S.
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23155 - 23162
  • [29] Detecting covert channels in cloud access control policies using Large Language Models
    Karmarkar, Hrishikesh
    Joshi, Vaibhavi
    Venkatesh, R.
    2024 IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND RESILIENCE, CSR, 2024, : 241 - 246
  • [30] Detecting Homophobic Speech in Soccer Tweets Using Large Language Models and Explainable AI
    Santos, Guto Leoni
    dos Santos, Vitor Gaboardi
    Kearns, Colm
    Sinclair, Gary
    Black, Jack
    Doidge, Mark
    Fletcher, Thomas
    Kilvington, Dan
    Liston, Katie
    Endo, Patricia Takako
    Lynn, Theo
    SOCIAL NETWORKS ANALYSIS AND MINING, ASONAM 2024, PT I, 2025, 15211 : 489 - 504