Improving ethical sensitivity for ethical decision-making in conversational artificial intelligence

被引:0
|
作者
Kyungsun Yoo [1 ]
Seongjin Ahn [1 ]
机构
[1] Sungkyunkwan University,Department of Computer Education
关键词
Ethics of artificial intelligence; Conversational artificial intelligence; Large language model; Ethical judgment; Ethical sensitivity;
D O I
10.1007/s10791-025-09520-6
中图分类号
学科分类号
摘要
The development of large language models has significantly advanced the reasoning capabilities of artificial intelligence (AI), often surpassing human-level performance. As AI's cognitive abilities rapidly progress, ethical concerns surrounding its applications have also increased. This indicates a heightened risk of bias, as AI models scale up and train on vast amounts of general data that inherently include social conventions related to gender, race, politics, and religion. This study proposes methods for enhancing ethical sensitivity to social bias. To achieve this, we defined 20 categories of social bias and developed a model that predicts the ethical sensitivity of sentences by leveraging the influence scores of words within these categories. The ethical sensitivity prediction model was validated using a paired-sample t-test to compare the ethical sensitivity evaluations of 25 AI-generated responses assessed by both AI and human evaluators. The test revealed no significant differences between the two groups, thus confirming the validity of the model. The findings of this study suggest that recognizing and predicting the ethical sensitivity of utterances concerning social biases can enhance ethical sensitivity, mitigate the risk of bias, and contribute to more ethical decision-making in AI interactions.
引用
收藏
相关论文
共 50 条