The Value of Trustworthy AI

被引:12
|
作者
Danks, David [1 ,2 ]
机构
[1] Carnegie Mellon Univ, Philosophy & Psychol, Pittsburgh, PA 15213 USA
[2] Carnegie Mellon Univ, Dept Philosophy, Pittsburgh, PA 15213 USA
关键词
Trust; Autonomous system; Psychology; Philosophy;
D O I
10.1145/3306618.3314228
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trust is one of the most critical relations in our human lives, whether trust in one another, trust in the artifacts that we use everyday, or trust of an AI system. Even a cursory examination of the literatures in human-computer interaction, human-robot interaction, and numerous other disciplines reveals a deep, persistent concern with the nature of trust in AI, and the conditions under which it can be generated, reduced, repaired, or influenced. At a high level, we often understand trust as a relation in which the trustor makes oneself vulnerable based on positive expectations about the behavior or intentions of the trustee [1]. For example, when I trust my car to start in the morning, I make myself vulnerable (e.g., I risk that I will be late to work if it does not start) because I have the positive expectation that it actually will start. This high-level characterization is relatively unhelpful, however, particularly given the wide range of disciplines that have examined the relation of trust, ranging from organizational behavior to game theory to ethics to cognitive science. The picture that emerges from, for example, social psychology (i.e., two distinct kinds of trust depending on whether one knows the trustee's behaviors or intentions/values) appears to be quite different from the one that emerges from moral philosophy (i.e., a single, highly-moralized notion), even though both are consistent with this high-level characterization. This talk first introduces that diversity of types of 'trust', but then argues that we can make progress towards a unified characterization by focusing on the function of trust. That is, we should ask why care whether we can trust our artifacts, AI, or fellow humans, as that can help to illuminate features of trust that are shared across domains, trustors, and trustees. I contend that one reason to desire trust is an "almost-necessary" condition on ethical action: namely, that the user has a reasonable belief that the system (whether human or machine) will behave approximately as intended. This condition is obviously not sufficient for ethical use, nor is it strictly necessary since the best available option might nonetheless be one for which the user lacks appropriate reasonable beliefs. Nonetheless, it provides a reasonable starting point for an analysis of 'trust'. More precisely, I propose that this condition indicates a role for trust as providing precisely those reasonable beliefs, at least when we have appropriately grounded trust. That is, we can understand 'appropriate trust' as obtaining when the trustor has justified beliefs that the trustee has suitable dispositions. As there is variation in the trustor's goals and values, and also the openness of the context of use, then different specific versions of 'appropriate trust' result as those variations lead to different types of focal dispositions, specific dispositions, or observability of dispositions, respectively. For example, in an open context (i.e., one where the possibilities cannot be exhaustively enumerated), the trustee's full dispositions will not be directly observable, but rather must be inferred from observations. This framework provides a unification of the different theories of 'trust' developed in different disciplines. Moreover, it provides clarity about one key function of trust, and thereby helps us to understand the value of (appropriate) trust. We need to trust our AI systems because that is a precondition for the ethical, responsible use of them.
引用
收藏
页码:521 / 522
页数:2
相关论文
共 50 条
  • [1] Trustworthy AI
    Wing, Jeannette M.
    [J]. COMMUNICATIONS OF THE ACM, 2021, 64 (10) : 64 - 71
  • [2] Trustworthy AI
    Singh, Richa
    Vatsa, Mayank
    Ratha, Nalini
    [J]. CODS-COMAD 2021: PROCEEDINGS OF THE 3RD ACM INDIA JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE & MANAGEMENT OF DATA (8TH ACM IKDD CODS & 26TH COMAD), 2021, : 449 - 453
  • [3] Trustworthy and responsible AI
    Eriksen, Remi
    [J]. Operations Engineer, 2024, (01): : 24 - 25
  • [4] Trustworthy AI for the People?
    Figueras, Claudia
    Verhagen, Harko
    Pargman, Teresa Cerratto
    [J]. AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 269 - 270
  • [5] How empty is Trustworthy AI? A discourse analysis of the Ethics Guidelines of Trustworthy AI
    Stamboliev, Eugenia
    Christiaens, Tim
    [J]. CRITICAL POLICY STUDIES, 2024,
  • [6] Network of AI and trustworthy: response to Simion and Kelp’s account of trustworthy AI
    Song F.
    [J]. Asian Journal of Philosophy, 2 (2):
  • [7] Trustworthy AI: AI made in Germany and Europe?
    Hirsch-Kreinsen, Hartmut
    Krokowski, Thorben
    [J]. AI & SOCIETY, 2023,
  • [8] An AI Harms and Governance Framework for Trustworthy AI
    Peckham, Jeremy B.
    [J]. COMPUTER, 2024, 57 (03) : 59 - 68
  • [9] The value of convergence research for developing trustworthy AI for weather, climate, and ocean hazards
    Amy McGovern
    Julie Demuth
    Ann Bostrom
    Christopher D. Wirz
    Philippe E. Tissot
    Mariana G. Cains
    Kate D. Musgrave
    [J]. npj Natural Hazards, 1 (1):
  • [10] Editorial: Trustworthy AI for healthcare
    Agafonov, Oleg
    Babic, Aleksandar
    Sousa, Sonia
    Alagaratnam, Sharmini
    [J]. FRONTIERS IN DIGITAL HEALTH, 2024, 6