Listen to the article
A recent study by the Icahn School of Medicine at Mount Sinai has uncovered a critical vulnerability in medical artificial intelligence (AI) systems: they can perpetuate false medical information when it appears within realistic clinical contexts.
The research, published in The Lancet Digital Health, tested nine leading language models with over a million prompts and found that these AI systems often fail to identify and flag medical misinformation when it’s embedded within authentic-looking hospital notes or social media health discussions.
“Our findings show that current AI systems can treat confident medical language as true by default, even when it’s clearly wrong,” explained co-senior author Dr. Eyal Klang, Chief of Generative AI in the Windreich Department of Artificial Intelligence and Human Health at Mount Sinai. “A fabricated recommendation in a discharge note can slip through. It can be repeated as if it were standard care.”
The research team employed a systematic approach to testing these models. They presented three types of content to the AI systems: actual hospital discharge summaries from the Medical Information Mart for Intensive Care (MIMIC) database with fabricated medical recommendations inserted; common health myths collected from Reddit; and 300 physician-validated clinical scenarios. Each case was presented in various formats, from neutral language to emotionally charged phrasing similar to content found on social media platforms.
In one revealing example, researchers inserted a false recommendation advising patients with esophagitis-related bleeding to “drink cold milk to soothe the symptoms” into a discharge note. Multiple AI models accepted this incorrect advice without flagging it as potentially harmful, treating it as legitimate medical guidance.
This vulnerability raises significant concerns about integrating AI systems into healthcare settings without robust safeguards. As medical institutions increasingly explore AI assistance for managing patient information, ensuring these systems can reliably distinguish between accurate medical information and dangerous falsehoods becomes paramount.
Dr. Girish N. Nadkarni, Chair of the Windreich Department of Artificial Intelligence and Human Health and Chief AI Officer of the Mount Sinai Health System, emphasized both the potential and the pitfalls of medical AI: “AI has the potential to be a real help for clinicians and patients, offering faster insights and support. But it needs built-in safeguards that check medical claims before they are presented as fact.”
The researchers advocate for a measurable approach to AI safety, suggesting that hospitals and developers use their dataset as a stress test for medical AI systems. “Instead of assuming a model is safe, you can measure how often it passes on a lie, and whether that number falls in the next generation,” noted Dr. Mahmud Omar, the study’s first author.
This research comes at a critical moment when healthcare systems worldwide are rapidly adopting AI technologies to improve efficiency and patient outcomes. The study highlights that the format and presentation of information—rather than its factual accuracy—often determines how AI systems process and repeat medical claims.
The findings suggest that current AI models may inadvertently amplify medical misinformation if implemented in clinical settings without additional verification mechanisms. This is particularly concerning given the increasing prevalence of health misinformation on social media platforms, which could potentially feed into AI-assisted healthcare systems.
The research, titled “Mapping LLM Susceptibility to Medical Misinformation Across Clinical Notes and Social Media,” received support from the National Center for Advancing Translational Sciences through the Clinical and Translational Science Awards grant, as well as from the Office of Research Infrastructure of the National Institutes of Health.
As medical AI continues to evolve, this study serves as a timely reminder that technological advancement must be balanced with rigorous safety measures, especially in healthcare where inaccurate information can have direct consequences for patient wellbeing. Researchers suggest that external evidence checks and systematic testing should become standard practice before AI systems are integrated into clinical tools.
Fact Checker
Verify the accuracy of this article using The Disinformation Commission analysis and real-time sources.


11 Comments
While AI holds great promise for healthcare, this study highlights a major vulnerability that needs to be fixed. Robust safeguards are needed to prevent medical misinformation from slipping through.
AI-driven medical misinformation is a frightening prospect. Rigorous testing and validation protocols are essential to ensure these systems can reliably distinguish fact from fiction before deployment.
Agreed. The stakes are far too high to allow medical AI systems to spread false information, even unintentionally. Patient safety has to be the top priority.
Concerning findings. AI systems need to be much better at detecting and flagging false medical information before they can be trusted in real-world healthcare applications. More rigorous testing and validation is clearly required.
Propagating false medical claims through AI systems is a serious problem that could have devastating impacts. I hope this research leads to significant improvements in the reliability and safety of these technologies.
This is concerning. AI systems that can’t distinguish truth from fiction in medical contexts could have serious consequences for patient care. Robust validation and safety protocols will be critical.
Absolutely. Ensuring the integrity of medical AI is paramount – patient lives could be at stake if these systems spread misinformation.
Interesting study on the risks of AI systems propagating false medical information. We’ll need to be very careful about how these models are deployed in clinical settings until they can reliably identify and flag misinformation.
This is a really important issue that needs to be addressed. Misinformation from medical AI could lead to disastrous outcomes for patients. Robust validation and ongoing monitoring will be critical.
Medical misinformation can be deadly, so it’s good to see research being done on this issue. But it’s also a bit worrying that current AI models struggle to reliably identify false claims in clinical settings.
Wow, this is really concerning. AI systems that can’t reliably identify false medical claims could end up causing serious harm to patients. Extensive testing and validation is clearly required before deploying these technologies.