The Value of Trustworthy AI

被引:12
|
作者
Danks, David [1 ,2 ]
机构
[1] Carnegie Mellon Univ, Philosophy & Psychol, Pittsburgh, PA 15213 USA
[2] Carnegie Mellon Univ, Dept Philosophy, Pittsburgh, PA 15213 USA
关键词
Trust; Autonomous system; Psychology; Philosophy;
D O I
10.1145/3306618.3314228
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trust is one of the most critical relations in our human lives, whether trust in one another, trust in the artifacts that we use everyday, or trust of an AI system. Even a cursory examination of the literatures in human-computer interaction, human-robot interaction, and numerous other disciplines reveals a deep, persistent concern with the nature of trust in AI, and the conditions under which it can be generated, reduced, repaired, or influenced. At a high level, we often understand trust as a relation in which the trustor makes oneself vulnerable based on positive expectations about the behavior or intentions of the trustee [1]. For example, when I trust my car to start in the morning, I make myself vulnerable (e.g., I risk that I will be late to work if it does not start) because I have the positive expectation that it actually will start. This high-level characterization is relatively unhelpful, however, particularly given the wide range of disciplines that have examined the relation of trust, ranging from organizational behavior to game theory to ethics to cognitive science. The picture that emerges from, for example, social psychology (i.e., two distinct kinds of trust depending on whether one knows the trustee's behaviors or intentions/values) appears to be quite different from the one that emerges from moral philosophy (i.e., a single, highly-moralized notion), even though both are consistent with this high-level characterization. This talk first introduces that diversity of types of 'trust', but then argues that we can make progress towards a unified characterization by focusing on the function of trust. That is, we should ask why care whether we can trust our artifacts, AI, or fellow humans, as that can help to illuminate features of trust that are shared across domains, trustors, and trustees. I contend that one reason to desire trust is an "almost-necessary" condition on ethical action: namely, that the user has a reasonable belief that the system (whether human or machine) will behave approximately as intended. This condition is obviously not sufficient for ethical use, nor is it strictly necessary since the best available option might nonetheless be one for which the user lacks appropriate reasonable beliefs. Nonetheless, it provides a reasonable starting point for an analysis of 'trust'. More precisely, I propose that this condition indicates a role for trust as providing precisely those reasonable beliefs, at least when we have appropriately grounded trust. That is, we can understand 'appropriate trust' as obtaining when the trustor has justified beliefs that the trustee has suitable dispositions. As there is variation in the trustor's goals and values, and also the openness of the context of use, then different specific versions of 'appropriate trust' result as those variations lead to different types of focal dispositions, specific dispositions, or observability of dispositions, respectively. For example, in an open context (i.e., one where the possibilities cannot be exhaustively enumerated), the trustee's full dispositions will not be directly observable, but rather must be inferred from observations. This framework provides a unification of the different theories of 'trust' developed in different disciplines. Moreover, it provides clarity about one key function of trust, and thereby helps us to understand the value of (appropriate) trust. We need to trust our AI systems because that is a precondition for the ethical, responsible use of them.
引用
收藏
页码:521 / 522
页数:2
相关论文
共 50 条
  • [21] Trustworthy AI for safe medicines
    Stegmann, Jens-Ulrich
    Littlebury, Rory
    Trengove, Markus
    Goetz, Lea
    Bate, Andrew
    Branson, Kim M.
    [J]. NATURE REVIEWS DRUG DISCOVERY, 2023, 22 (10) : 855 - 856
  • [22] Towards Trustworthy AI in Dentistry
    Ma, J.
    Schneider, L.
    Lapuschkin, S.
    Achtibat, R.
    Duchrau, M.
    Krois, J.
    Schwendicke, F.
    Samek, W.
    [J]. JOURNAL OF DENTAL RESEARCH, 2022, 101 (11) : 1263 - 1268
  • [23] Trustworthy AI: A Computational Perspective
    Liu, Haochen
    Wang, Yiqi
    Fan, Wenqi
    Liu, Xiaorui
    Li, Yaxin
    Jain, Shaili
    Liu, Yunhao
    Jain, Anil
    Tang, Jiliang
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (01)
  • [24] Guest Editorial: Trustworthy AI
    Jin, Yier
    Ho, Tsung-Yi
    [J]. ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (03)
  • [25] Trustworthy AI'21: the 1st International Workshop on Trustworthy AI for Multimedia Computing
    Furon, Teddy
    Liu, Jingen
    Rawat, Yogesh
    Zhang, Wei
    Zhao, Qi
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 5708 - 5709
  • [26] Trustworthy AI Means Public AI [Last Word]
    Schneier, Bruce
    [J]. IEEE SECURITY & PRIVACY, 2023, 21 (06) : 95 - 96
  • [27] AI Augmentation for Trustworthy AI: Augmented Robot Teleoperation
    Marino, Daniel L.
    Grandio, Javier
    Wickramasinghe, Chathurika S.
    Schroeder, Kyle
    Bourne, Keith
    Filippas, Afroditi, V
    Manic, Milos
    [J]. 2020 13TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2020, : 155 - 161
  • [28] Filling gaps in trustworthy development of AIFilling gaps in trustworthy development of AI
    Avin, Shahar
    Belfield, Haydn
    Brundage, Miles
    Krueger, Gretchen
    Wang, Jasmine
    Weller, Adrian
    Anderljung, Markus
    Krawczuk, Igor
    Krueger, David
    Lebensold, Jonathan
    Maharaj, Tegan
    Zilberman, Noa
    [J]. SCIENCE, 2021, 374 (6573) : 1327 - 1329
  • [29] Trustworthy AI-Part II
    Mariani, Riccardo
    Rossi, Francesca
    Cucchiara, Rita
    Pavone, Marco
    Simkin, Barnaby
    Koene, Ansgar
    Papenbrock, Jochen
    [J]. COMPUTER, 2023, 56 (05) : 13 - 16
  • [30] Trustworthy AI and the Logics of Intersectional Resistance
    Knowles, Bran
    Fledderjohann, Jasmine
    Richards, John T.
    Varshney, Kush R.
    [J]. PROCEEDINGS OF THE 6TH ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2023, 2023, : 172 - 182