Exploring the risks of automation bias in healthcare artificial intelligence applications: A Bowtie analysis

This study conducts an in-depth review and Bowtie analysis of automation bias in AI-driven Clinical Decision Support Systems (CDSSs) within healthcare settings. Automation bias, the tendency of human operators to over-rely on automated systems, poses a critical challenge in implementing AI-driven te...

Celý popis

Uložené v:
Podrobná bibliografia
Vydané v:Journal of Safety Science and Resilience = An quan ke xue yu ren xing (Ying wen) Ročník 5; číslo 4; s. 460 - 469
Hlavní autori: Abdelwanis, Moustafa, Alarafati, Hamdan Khalaf, Tammam, Maram Muhanad Saleh, Simsekler, Mecit Can Emre
Médium: Journal Article
Jazyk:English
Vydavateľské údaje: Elsevier B.V 01.12.2024
KeAi Communications Co., Ltd
Predmet:
ISSN:2666-4496, 2666-4496
On-line prístup:Získať plný text
Tagy: Pridať tag
Žiadne tagy, Buďte prvý, kto otaguje tento záznam!
Popis
Shrnutí:This study conducts an in-depth review and Bowtie analysis of automation bias in AI-driven Clinical Decision Support Systems (CDSSs) within healthcare settings. Automation bias, the tendency of human operators to over-rely on automated systems, poses a critical challenge in implementing AI-driven technologies. To address this challenge, Bowtie analysis is employed to examine the causes and consequences of automation bias affected by over-reliance on AI-driven systems in healthcare. Furthermore, this study proposes preventive measures to address automation bias during the design phase of AI model development for CDSSs, along with effective mitigation strategies post-deployment. The findings highlight the imperative role of a systems approach, integrating technological advancements, regulatory frameworks, and collaborative endeavors between AI developers and healthcare practitioners to diminish automation bias in AI-driven CDSSs. We further identify future research directions, proposing quantitative evaluations of the mitigation and preventative measures.
ISSN:2666-4496
2666-4496
DOI:10.1016/j.jnlssr.2024.06.001