A trustworthy AI reality-check: the lack of transparency of artificial intelligence products in healthcare

被引:3
|
作者
Fehr, Jana [1 ,2 ,3 ]
Citro, Brian [4 ]
Malpani, Rohit [5 ]
Lippert, Christoph [1 ,2 ,4 ]
Madai, Vince I. [3 ,5 ]
机构
[1] Hasso Plattner Inst, Digital Hlth & Machine Learning, Potsdam, Germany
[2] Univ Potsdam, Digital Engn Fac, Potsdam, Germany
[3] Charite Univ Med Berlin, Berlin Inst Hlth BIH, QUEST Ctr Responsible Res, Berlin, Germany
[4] Icahn Sch Med Mt Sinai, Hasso Plattner Inst Digital Hlth Mt Sinai, New York, NY USA
[5] Birmingham City Univ, Fac Comp Engn & Built Environm, Sch Comp & Digital Technol, Birmingham, England
来源
关键词
medical AI; AI ethics; transparency; medical device regulation; trustworthy AI; MEDICAL DEVICES; BIAS;
D O I
10.3389/fdgth.2024.1267290
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Trustworthy medical AI requires transparency about the development and testing of underlying algorithms to identify biases and communicate potential risks of harm. Abundant guidance exists on how to achieve transparency for medical AI products, but it is unclear whether publicly available information adequately informs about their risks. To assess this, we retrieved public documentation on the 14 available CE-certified AI-based radiology products of the II b risk category in the EU from vendor websites, scientific publications, and the European EUDAMED database. Using a self-designed survey, we reported on their development, validation, ethical considerations, and deployment caveats, according to trustworthy AI guidelines. We scored each question with either 0, 0.5, or 1, to rate if the required information was "unavailable", "partially available," or "fully available." The transparency of each product was calculated relative to all 55 questions. Transparency scores ranged from 6.4% to 60.9%, with a median of 29.1%. Major transparency gaps included missing documentation on training data, ethical considerations, and limitations for deployment. Ethical aspects like consent, safety monitoring, and GDPR-compliance were rarely documented. Furthermore, deployment caveats for different demographics and medical settings were scarce. In conclusion, public documentation of authorized medical AI products in Europe lacks sufficient public transparency to inform about safety and risks. We call on lawmakers and regulators to establish legally mandated requirements for public and substantive transparency to fulfill the promise of trustworthy AI for health.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Apocalyptic AI: Visions of Heaven in Robotics, Artificial Intelligence, and Virtual Reality
    Graziano, Michael
    [J]. JOURNAL OF RELIGION AND POPULAR CULTURE, 2014, 26 (02): : 267 - 268
  • [32] Connecting the dots in trustworthy Artificial Intelligence: From AI principles, ethics, and key requirements to responsible AI systems and regulation
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Coeckelbergh, Mark
    de Prado, Marcos Lopez
    Herrera-Viedma, Enrique
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2023, 99
  • [33] Development of a practical training method for a healthcare artificial intelligence (AI) chatbot
    Lin, Julia
    Joseph, Todd
    Parga-Belinkie, Joanna Jean
    Mandel, Abigail
    Schumacher, Ryan
    Neumann, Karl
    Scalise, Laura
    Gaulton, Jessica
    Christ, Lori
    Leitner, Kirstin
    Rosin, Roy
    [J]. BMJ INNOVATIONS, 2021, 7 (02) : 441 - 444
  • [34] Lay individuals' perceptions of artificial intelligence (AI)-empowered healthcare systems
    School of Computer Science and Information Systems, Pace University, New York
    NY, United States
    不详
    FL, United States
    不详
    MA, United States
    不详
    [J]. Proceedings of the Association for Information Science and Technology, 2020, 1
  • [35] RACIST ROBOTS AND THE LACK OF LEGAL REMEDIES IN THE USE OF ARTIFICIAL INTELLIGENCE IN HEALTHCARE
    Dueno, Tyler
    [J]. CONNECTICUT INSURANCE LAW JOURNAL, 2020, 27 (01): : 337 - 358
  • [36] Explainable artificial intelligence to increase transparency for revolutionizing healthcare ecosystem and the road ahead
    Roy, Sudipta
    Pal, Debojyoti
    Meena, Tanushree
    [J]. NETWORK MODELING AND ANALYSIS IN HEALTH INFORMATICS AND BIOINFORMATICS, 2023, 13 (01):
  • [37] Explainable artificial intelligence to increase transparency for revolutionizing healthcare ecosystem and the road ahead
    Sudipta Roy
    Debojyoti Pal
    Tanushree Meena
    [J]. Network Modeling Analysis in Health Informatics and Bioinformatics, 13
  • [38] Lack of Transparency and Potential Bias in Artificial Intelligence Data Sets and Algorithms A Scoping Review
    Daneshjou, Roxana
    Smith, Mary P.
    Sun, Mary D.
    Rotemberg, Veronica
    Zou, James
    [J]. JAMA DERMATOLOGY, 2021, 157 (11) : 1362 - 1369
  • [39] Do citizens trust trustworthy artificial intelligence? Experimental evidence on the limits of ethical AI measures in government
    Kleizen, Bjorn
    Van Dooren, Wouter
    Verhoest, Koen
    Tan, Evrim
    [J]. GOVERNMENT INFORMATION QUARTERLY, 2023, 40 (04)
  • [40] Applying the Assessment List for Trustworthy Artificial Intelligence on the development of AI supported Air Traffic Controller Operations
    Stefani, Thomas
    Deligiannaki, Fotini
    Berro, Charles
    Jameel, Mohsan
    Hunger, Robert
    Bruder, Carmen
    Krueger, Thomas
    [J]. 2023 IEEE/AIAA 42ND DIGITAL AVIONICS SYSTEMS CONFERENCE, DASC, 2023,