Tampering with Generative Artificial Intelligence by Jailbreaking

被引:0
|
作者
Claverini, Corrado [1 ]
机构
[1] Univ Salento, Lecce, Italy
来源
TEORIA-RIVISTA DI FILOSOFIA | 2024年 / 44卷 / 01期
关键词
ChatGPT; ethics of artificial intelligence; generative artificial in- telligence; jailbreaking; regulation of artificial intelligence;
D O I
10.4454/mg6wax06
中图分类号
B [哲学、宗教];
学科分类号
01 ; 0101 ;
摘要
In this paper, I will analyse the risks linked to the use of generative artificial intelligence systems and relative risk-reduction strategies, while concentrating in particular on the possibility of tampering with the chatbot ChatGPT by jailbreaking. After examining how a user can tamper with this generative AI, bypassing its ethical and legal restrictions, through a series of prompts, I will turn my focus to the ethical issues raised by the malicious use of this technology: are the transparency requirements requested of generative AI sufficient or should there be tighter restrictions that do not hinder the innovation and development of these technologies? How can the risk of tampering with these AI tools be lowered? And, should a breach take place, who is responsible: the AI developer or the jailbreaker? To what extent could the changes needed to prevent jailbreaking involuntarily generate or strengthen certain biases? In conclusion, I will uphold the necessity of ethical reflection for the sustainable and "human-centric" development of AI.
引用
收藏
页数:172
相关论文
共 50 条