• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to secondary sidebar
  • Skip to footer

  • Opinion
  • Health IT
    • Behavioral Health
    • Care Coordination
    • EMR/EHR
    • Interoperability
    • Patient Engagement
    • Population Health Management
    • Revenue Cycle Management
    • Social Determinants of Health
  • Digital Health
    • AI
    • Blockchain
    • Precision Medicine
    • Telehealth
    • Wearables
  • Life Sciences
  • Investments
  • M&A
  • Value-based Care
    • Accountable Care (ACOs)
    • Medicare Advantage

Mount Sinai Study: LLMs Susceptible to Medical Misinformation in Clinical Notes

by Fred Pennic 02/09/2026 Leave a Comment

  • LinkedIn
  • Twitter
  • Facebook
  • Email
  • Print
Icahn Mount Sinai Becomes First US Medical School to Provide ChatGPT Edu Access to All Students

What You Should Know

  • The Study: In a paper published today in The Lancet Digital Health [10.1016/j.landig.2025.100949], researchers at the Icahn School of Medicine at Mount Sinai analyzed over one million prompts across nine leading Large Language Models (LLMs) to test their susceptibility to medical misinformation.
  • The Vulnerability: The study found that AI models frequently repeat false medical claims—such as advising patients with bleeding to “drink cold milk”—if the lie is embedded in realistic hospital notes or professional-sounding language.
  • The Takeaway: Current safeguards are failing to distinguish fact from fiction when the fiction “sounds” like a doctor. For these models, the style of the writing (confident, clinical) often overrides the truth of the content.

The “Cold Milk” Fallacy

To test the systems, the research team exposed nine leading LLMs to over one million prompts. They took real hospital discharge summaries (from the MIMIC database) and injected them with single, fabricated recommendations.

The results were sobering. In one specific example, a discharge note for a patient with esophagitis-related bleeding falsely advised them to “drink cold milk to soothe the symptoms”—a recommendation that is clinically unsafe.

Instead of flagging this as dangerous, several models accepted the statement as fact. They processed it, repeated it, and treated it like ordinary medical guidance simply because it appeared in a format that looked like a valid hospital note.

Style Over Substance

“Our findings show that current AI systems can treat confident medical language as true by default, even when it’s clearly wrong,” said Dr. Eyal Klang, Chief of Generative AI at Mount Sinai.

This exposes a fundamental flaw in how current LLMs operate in healthcare. They are not necessarily verifying the medical accuracy of a claim against a database of truth; they are predicting the next word based on context. If the context is a highly realistic, professional discharge summary, the model assumes the content within it is accurate.

“For these models, what matters is less whether a claim is correct than how it is written,” Klang added.

The “Stress Test” Solution

The implications for clinical deployment are massive. If an AI summarizer is used to condense patient records, and one of those records contains a human error (or a hallucination from a previous AI), the system might amplify that error rather than catch it.

Dr. Mahmud Omar, the study’s first author, argues that we need a new standard for validation. “Instead of assuming a model is safe, you can measure how often it passes on a lie,” he said. The authors propose using their dataset as a standard “stress test” for any medical AI before it is allowed near a patient.

  • LinkedIn
  • Twitter
  • Facebook
  • Email
  • Print

Tagged With: Artificial Intelligence

Tap Native

Get in-depth healthcare technology analysis and commentary delivered straight to your email weekly

Reader Interactions

Primary Sidebar

Subscribe to HIT Consultant

Latest insightful articles delivered straight to your inbox weekly.

Submit a Tip or Pitch

Featured Research Report

2026 Best in KLAS Awards: The Full List of Software & Services Winners

Most-Read

The "Platform" Squeeze: Epic Releases Native AI Charting, Putting Venture-Backed Scribes on Notice

The “Platform” Squeeze: Epic Releases Native AI Charting, Putting Venture-Backed Scribes on Notice

Analysis: Oracle Cerner’s Plans for a National EHR

Oracle May Cut 30k Jobs and Sell Cerner to Fund $156B OpenAI Deal

The $1.9B Exit: Why CommonSpirit is Insourcing Revenue Cycle and Tenet is Betting Big on Conifer AI

The $1.9B Exit: Why CommonSpirit is Insourcing Revenue Cycle and Tenet is Betting Big on Conifer AI

KLAS 2026 Rankings: Aledade and Guidehealth Named Top VBC Enablement Firms

KLAS 2026 Rankings: Aledade and Guidehealth Named Top VBC Enablement Firms

Beyond the Hype: New KLAS Data Validates the Financial and Clinical ROI of Ambient AI

Beyond the Hype: New KLAS Data Validates the Financial and Clinical ROI of Ambient AI

Anthropic Debuts ‘Claude for Healthcare’ and Opus 4.5 to Engineer the Future of Life Sciences

Anthropic Debuts ‘Claude for Healthcare’ and Opus 4.5 to Engineer the Future of Life Sciences

OpenAI Debuts ChatGPT Health: A ‘Digital Front Door’ That Connects Medical Records to Agentic AI

OpenAI Debuts ChatGPT Health: A ‘Digital Front Door’ That Connects Medical Records to Agentic AI

From Genes to Hackers: The Hidden Cybersecurity Risks in Life Sciences

From Genes to Hackers: The Hidden Cybersecurity Risks in Life Sciences

Utah Becomes First State to Approve AI System for Prescription Renewals

Utah Becomes First State to Approve AI System for Prescription Renewals

NYC Health + Hospitals to Acquire Maimonides in $2.2B Safety Net Overhaul

NYC Health + Hospitals to Acquire Maimonides in $2.2B Safety Net Overhaul

Secondary Sidebar

Footer

Company

  • About Us
  • 2026 Editorial Calendar
  • Advertise with Us
  • Reprints and Permissions
  • Op-Ed Submission Guidelines
  • Contact
  • Subscribe

Editorial Coverage

  • Opinion
  • Health IT
    • Care Coordination
    • EMR/EHR
    • Interoperability
    • Population Health Management
    • Revenue Cycle Management
  • Digital Health
    • Artificial Intelligence
    • Blockchain Tech
    • Precision Medicine
    • Telehealth
    • Wearables
  • Startups
  • Value-Based Care
    • Accountable Care
    • Medicare Advantage

Connect

Subscribe to HIT Consultant Media

Latest insightful articles delivered straight to your inbox weekly

Copyright © 2026. HIT Consultant Media. All Rights Reserved. Privacy Policy |