{"title":"The illusion of safety: A report to the FDA on AI healthcare product approvals.","authors":"Rawan Abulibdeh, Leo Anthony Celi, Ervin Sejdić","doi":"10.1371/journal.pdig.0000866","DOIUrl":null,"url":null,"abstract":"<p><p>Artificial intelligence is rapidly transforming healthcare, offering promising advancements in diagnosis, treatment, and patient outcomes. However, concerns regarding the regulatory oversight of artificial intelligence driven medical technologies have emerged, particularly with the U.S. Food and Drug Administration's current approval processes. This paper critically examines the U.S. Food and Drug Administration's regulatory framework for artificial intelligence powered healthcare products, highlighting gaps in safety evaluations, post-market surveillance, and ethical considerations. Artificial intelligence's continuous learning capabilities introduce unique risks, as algorithms evolve beyond their initial validation, potentially leading to performance degradation and biased outcomes. Although the U.S. Food and Drug Administration has taken steps to address these challenges, such as artificial intelligence/machine learning-based software as a medical device action plan and proposed regulatory adjustments, significant weaknesses remain, particularly in real-time monitoring, transparency and bias mitigation. This paper argues for a more adaptive, community-engaged regulatory approach that mandates extensive post-market evaluations, requires artificial intelligence developers to disclose training data sources, and establishes enforceable standards for fairness, equity, and accountability. A patient-centered regulatory framework must also integrate diverse perspectives to ensure artificial intelligence technologies serve all populations equitably. By fostering an agile, transparent, and ethics-driven oversight system, the U.S. Food and Drug Administration can balance innovation with patient safety, ensuring that artificial intelligence-driven medical technologies enhance, rather than compromise, healthcare outcomes.</p>","PeriodicalId":74465,"journal":{"name":"PLOS digital health","volume":"4 6","pages":"e0000866"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12140231/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"PLOS digital health","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1371/journal.pdig.0000866","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/6/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Artificial intelligence is rapidly transforming healthcare, offering promising advancements in diagnosis, treatment, and patient outcomes. However, concerns regarding the regulatory oversight of artificial intelligence driven medical technologies have emerged, particularly with the U.S. Food and Drug Administration's current approval processes. This paper critically examines the U.S. Food and Drug Administration's regulatory framework for artificial intelligence powered healthcare products, highlighting gaps in safety evaluations, post-market surveillance, and ethical considerations. Artificial intelligence's continuous learning capabilities introduce unique risks, as algorithms evolve beyond their initial validation, potentially leading to performance degradation and biased outcomes. Although the U.S. Food and Drug Administration has taken steps to address these challenges, such as artificial intelligence/machine learning-based software as a medical device action plan and proposed regulatory adjustments, significant weaknesses remain, particularly in real-time monitoring, transparency and bias mitigation. This paper argues for a more adaptive, community-engaged regulatory approach that mandates extensive post-market evaluations, requires artificial intelligence developers to disclose training data sources, and establishes enforceable standards for fairness, equity, and accountability. A patient-centered regulatory framework must also integrate diverse perspectives to ensure artificial intelligence technologies serve all populations equitably. By fostering an agile, transparent, and ethics-driven oversight system, the U.S. Food and Drug Administration can balance innovation with patient safety, ensuring that artificial intelligence-driven medical technologies enhance, rather than compromise, healthcare outcomes.