15 February, 2026
ai-in-healthcare-faces-data-poisoning-threats-study-warns

Artificial Intelligence (AI) has rapidly integrated into healthcare, enhancing diagnosis, documentation, triage, and treatment planning. However, a new study published in the Journal of Medical Internet Research by researcher Farhad Abtahi reveals a critical vulnerability in AI systems: data poisoning. This issue arises when manipulated training data leads AI to make unsafe decisions. The study emphasizes that while healthcare systems are not currently under attack, the potential for such risks warrants serious attention.

Abtahi’s research synthesizes existing security studies, illustrating how data poisoning could realistically occur within healthcare settings. The focus is on how compromised data may blend seamlessly into normal clinical workflows, making detection of these threats difficult. The key concern is that AI systems, which are often perceived as secure due to their reliance on extensive datasets, can still be vulnerable to attacks that target the training pipeline.

The paper highlights a widespread assumption that the presence of large datasets inherently dilutes the impact of bad data. Contrary to this belief, the findings suggest that the success of such attacks can depend significantly on the number of poisoned samples, not merely their proportion within the dataset. Previous studies indicate that as few as 100-500 poisoned samples can severely compromise AI systems, with attack success rates often exceeding 60%.

Understanding the Risks of Data Poisoning

One illustrative scenario presented in the paper is the “medical scribe Sybil attack,” a hypothetical model that demonstrates how coordinated actors could introduce poisoned data through legitimate clinical activities. In this scenario, actors arrange numerous legitimate-seeming patient visits with scripted histories. If an AI-powered scribe documents these visits, the resulting records may be used as training data for multiple downstream AI systems, leading to broader vulnerabilities.

Abtahi clarifies that these scenarios are not based on documented incidents but are designed to outline plausible pathways for data poisoning. The objective is to strengthen preventive measures before any real harm occurs. He stresses that there is no singular solution to completely eliminate the risk of data poisoning. Instead, he advocates for a comprehensive approach, referred to as “defense in depth,” which involves multiple protective layers across the system lifecycle.

The EU AI Act, with its risk-based structure, addresses some of the issues related to data poisoning. It mandates high-risk AI systems to adhere to stringent requirements concerning data governance, risk management, and ongoing performance monitoring. These regulations aim to simplify the detection and response process should incidents arise.

Challenges in Prevention and Detection

While the EU AI Act introduces beneficial frameworks, Abtahi notes that the subtle nature of data poisoning makes prevention particularly challenging. Problems may not materialize until months or years after the introduction of compromised data, potentially resulting in significant harm. Consequently, the study argues for integrated safeguards throughout the data pipeline, model development, and deployment phases, rather than relying solely on any one regulatory measure.

Abtahi concludes that proactive measures are essential to mitigate risks associated with data poisoning in AI healthcare applications. The ongoing development of effective strategies will be critical in ensuring the safety and efficacy of AI technologies in clinical practice. The comprehensive study titled “Data Poisoning Vulnerabilities Across Health Care Artificial Intelligence Architectures: Analytical Security Framework and Defense Strategies” serves as a valuable resource for understanding these emerging threats and the necessary steps to counter them.