Listen to the article

0:00
0:00

Medical AI Shows Vulnerability to Authoritative Health Misinformation, Mount Sinai Study Reveals

Medical artificial intelligence systems can perpetuate dangerous health misinformation, particularly when it comes from seemingly trustworthy sources presented in an authoritative manner, according to a comprehensive new study by researchers at Mount Sinai Health System.

The research, published in The Lancet Digital Health, found that large language models (LLMs) often fail to detect false medical claims when they’re couched in professional clinical language, similar to how many healthcare consumers might accept information from someone perceived as an expert.

“Our findings show that current AI systems can treat confident medical language as true by default, even when it’s clearly wrong,” explained Dr. Eyal Klang, chief of generative AI at the Icahn School of Medicine at Mount Sinai and co-senior author of the study. “A fabricated recommendation in a discharge note can slip through. It can be repeated as if it were standard care.”

For these AI systems, Klang emphasized, “what matters is less whether a claim is correct than how it is written.”

The Mount Sinai team conducted an extensive analysis, testing 20 large language models with over 3.4 million prompts containing health misinformation. They sourced these false claims from three realistic scenarios: social media conversations, actual hospital discharge notes with inserted false recommendations, and 300 physician-validated simulated clinical vignettes.

To examine how rhetorical presentation influences AI behavior, researchers presented each prompt in a neutral format and then ten additional times, each employing a different logical fallacy such as circular reasoning, hasty generalization, or straw man arguments. They then recorded whether the model accepted the false information or recognized the faulty reasoning.

The results revealed that LLMs were most vulnerable when encountering medical misinformation presented in authoritative clinical language. Surprisingly, the models became less susceptible when the same false claims were wrapped in recognizable logical fallacies, suggesting that improving AI safety may depend more on context-awareness and fact-grounding than on increasing model size.

This vulnerability mirrors human behavior in many ways. Just as patients might accept questionable medical advice from someone in a white coat using technical terminology, AI systems can be swayed by authoritative presentation rather than factual accuracy.

In an accompanying commentary published in the same journal, psychology professors Dr. Sander van der Linden and Yara Kyrychenko from the University of Cambridge suggested that LLMs need to be “immunized” against medical misinformation through specialized training.

“Inoculation prompting during training can be an effective method for preventing misaligned model behavior in the future,” they wrote. This approach involves deliberately exposing AI systems to misinformation in controlled settings, helping them learn to recognize and reject false claims.

“Adding a system prompt that instructs an LLM to produce misinformation and fine-tuning on a dataset of false healthcare-related claims could increase the model’s understanding of what health misinformation is,” the Cambridge researchers explained. “When prompted to be helpful instead of producing misinformation, the model should be more likely to generate truthful responses or push back on false claims.”

The good news emerging from the research is that AI can be trained to recognize logical fallacies and alert users when shaky reasoning is detected. Dr. Mahmud Omar, co-lead author of the study, suggested that healthcare organizations and AI developers could use the research dataset as a “stress test” for medical AI systems.

“Instead of assuming a model is safe, you can measure how often it passes along a lie, and whether that number falls in the next generation,” Omar noted.

The findings come as healthcare systems increasingly explore AI implementation for clinical documentation, patient communication, and decision support. The study underscores the importance of developing robust safeguards and evaluation frameworks before deploying these technologies in sensitive medical contexts.

Both the Mount Sinai study and the Cambridge commentary are available in full for free, providing valuable resources for researchers, developers, and healthcare organizations working to ensure responsible AI development in medicine.

Fact Checker

Verify the accuracy of this article using The Disinformation Commission analysis and real-time sources.

16 Comments

  1. The vulnerability of medical AI to authoritative-sounding misinformation is a critical issue that deserves close attention. Ensuring these systems can reliably distinguish fact from fiction, especially in sensitive health domains, should be a top priority for the research community.

    • Agreed. The stakes are too high to allow medical AI systems to spread harmful falsehoods, even inadvertently. Robust safeguards and discernment capabilities must be built in from the ground up.

  2. This study underscores the importance of developing medical AI with strong safeguards against perpetuating misinformation. The tendency to treat confident clinical language as true by default, even when incorrect, is a serious concern that needs to be addressed.

    • Michael Williams on

      You’re right, this is a significant challenge that must be overcome. Rigorous testing and validation processes will be essential to ensure these systems can reliably distinguish fact from fiction in the medical domain.

  3. This study is a sobering reminder of the challenges involved in developing safe and reliable medical AI systems. The vulnerability to authoritative-sounding misinformation is a serious issue that requires the full attention and creative problem-solving of the research community.

  4. Oliver S. Rodriguez on

    This study highlights the need for extreme caution when deploying medical AI systems. The potential for perpetuating dangerous misinformation is alarming and must be taken very seriously by researchers and developers. Rigorous testing and validation protocols are essential.

  5. The findings here are concerning, but not entirely surprising. Large language models can struggle with nuance and contextual understanding, which makes them vulnerable to being misled by authoritative-sounding but inaccurate medical information. Addressing this will require innovative approaches.

    • Absolutely. Developing more advanced discernment capabilities, perhaps through techniques like few-shot learning or meta-learning, could help these models better detect and filter out misleading claims, even when they’re couched in convincing language.

  6. This study highlights the critical importance of developing medical AI systems with strong safeguards against the perpetuation of misinformation. The tendency to accept authoritative-sounding claims as true, even when they’re false, is a serious concern that must be addressed head-on.

  7. The findings from Mount Sinai are concerning but not entirely unexpected. The inherent limitations of current AI systems when it comes to nuanced reasoning and contextual understanding make them vulnerable to being misled by convincing but inaccurate medical information. Addressing this is crucial.

    • Elizabeth Thompson on

      You’re right, this is a complex challenge that will require a multifaceted approach. Improving discernment capabilities, enhancing safety and quality controls, and developing more robust validation protocols will all be essential in mitigating the risks of medical AI misinformation.

  8. Jennifer Williams on

    The findings from this Mount Sinai study are deeply concerning. The potential for large language models to inadvertently spread dangerous medical misinformation is alarming and underscores the need for rigorous testing and validation protocols. Responsible development of these systems is paramount.

    • Absolutely. Ensuring medical AI can reliably distinguish fact from fiction, even when claims are presented in a convincing manner, will be essential. Innovative approaches to discernment and contextual understanding will be key to mitigating these risks.

  9. Oliver Thompson on

    This study is a wake-up call for the AI research community. The tendency of large language models to accept authoritative-sounding medical claims as true, even when they’re false, is a serious problem that needs immediate attention and innovative solutions.

  10. John T. Martinez on

    Interesting study on the vulnerability of medical AI to misinformation. It highlights the need for robust safety and quality controls, as well as advanced discernment capabilities, to prevent the spread of dangerous falsehoods. Responsible development of these systems is crucial.

    • Agreed. Ensuring AI models can accurately identify and filter out misleading or inaccurate medical claims is critical, especially when they’re presented in an authoritative manner. Proper training and validation protocols will be key.

Leave A Reply

A professional organisation dedicated to combating disinformation through cutting-edge research, advanced monitoring tools, and coordinated response strategies.

Company

Disinformation Commission LLC
30 N Gould ST STE R
Sheridan, WY 82801
USA

© 2026 Disinformation Commission LLC. All rights reserved.