Enabling collaborative governance of medical AI
Publikation: Bidrag til tidsskrift › Tidsskriftartikel › Forskning › fagfællebedømt
Standard
Enabling collaborative governance of medical AI. / Price, W. Nicholson; Sendak, Mark; Balu, Suresh; Singh, Karandeep.
I: Nature Machine Intelligence, Bind 5, Nr. 8, 2024, s. 821-823.Publikation: Bidrag til tidsskrift › Tidsskriftartikel › Forskning › fagfællebedømt
Harvard
APA
Vancouver
Author
Bibtex
}
RIS
TY - JOUR
T1 - Enabling collaborative governance of medical AI
AU - Price, W. Nicholson
AU - Sendak, Mark
AU - Balu, Suresh
AU - Singh, Karandeep
PY - 2024
Y1 - 2024
N2 - Artificial intelligence (AI) is rapidly entering healthcare, from sepsis prediction to image analysis to patient management. Some AI systems are developed by venture-backed start-ups, others are homegrown, and many are embedded within electronic health records (EHR) systems. They demand governance: the task of ensuring safety and effectiveness at the time of integration into clinical care and throughout the product lifecycle. AI systems, including broadly deployed systems, have shown substantial quality problems and implementation challenges despite their overall promise1. Our team includes leaders at Michigan Medicine and Duke Health with substantial on-the-ground experience developing, implementing and maintaining AI systems used in clinical practice and extensive experience supporting government actors seeking to scale the potential benefits of AI. We argue the inadequacy of an exclusive focus on centralized governance — by, for example, the Food and Drug Administration (FDA), the Office of the National Coordinator for Health Information Technology (ONC), the Centers for Medicare and Medicaid Services (CMS) or even the Federal Trade Commission. Instead, centralized governors must also coordinate and support local governance within healthcare delivery settings with varying resources and capabilities in a model of collaborative governance.
AB - Artificial intelligence (AI) is rapidly entering healthcare, from sepsis prediction to image analysis to patient management. Some AI systems are developed by venture-backed start-ups, others are homegrown, and many are embedded within electronic health records (EHR) systems. They demand governance: the task of ensuring safety and effectiveness at the time of integration into clinical care and throughout the product lifecycle. AI systems, including broadly deployed systems, have shown substantial quality problems and implementation challenges despite their overall promise1. Our team includes leaders at Michigan Medicine and Duke Health with substantial on-the-ground experience developing, implementing and maintaining AI systems used in clinical practice and extensive experience supporting government actors seeking to scale the potential benefits of AI. We argue the inadequacy of an exclusive focus on centralized governance — by, for example, the Food and Drug Administration (FDA), the Office of the National Coordinator for Health Information Technology (ONC), the Centers for Medicare and Medicaid Services (CMS) or even the Federal Trade Commission. Instead, centralized governors must also coordinate and support local governance within healthcare delivery settings with varying resources and capabilities in a model of collaborative governance.
U2 - 10.1038/s42256-023-00699-1
DO - 10.1038/s42256-023-00699-1
M3 - Journal article
VL - 5
SP - 821
EP - 823
JO - Nature Machine Intelligence
JF - Nature Machine Intelligence
SN - 2522-5839
IS - 8
ER -
ID: 389363271