• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to secondary sidebar
  • Skip to footer

  • Opinion
  • Health IT
    • Behavioral Health
    • Care Coordination
    • EMR/EHR
    • Interoperability
    • Patient Engagement
    • Population Health Management
    • Revenue Cycle Management
    • Social Determinants of Health
  • Digital Health
    • AI
    • Blockchain
    • Precision Medicine
    • Telehealth
    • Wearables
  • Life Sciences
  • Investments
  • M&A
  • Value-based Care
    • Accountable Care (ACOs)
    • Medicare Advantage

AI Chatbots Easily Misled by False Medical Information, Mount Sinai Study Finds

by Fred Pennic 08/06/2025 Leave a Comment

  • LinkedIn
  • Twitter
  • Facebook
  • Email
  • Print
Icahn Mount Sinai Becomes First US Medical School to Provide ChatGPT Edu Access to All Students

What You Should Know: 

– A new study from the Icahn School of Medicine at Mount Sinai reveals that leading AI chatbots are highly susceptible to repeating and even elaborating on false medical information. 

– However, the researchers also demonstrated a practical solution: a simple, one-line warning prompt can significantly reduce this risk, highlighting a crucial path forward for safely integrating these tools into health care. The findings were published in the August 2 issue of Communications Medicine.

AI Vulnerability Leads to Medical Misinformation

The study, titled “Large Language Models Demonstrate Widespread Hallucinations for Clinical Decision Support: A Multiple Model Assurance Analysis,” underscores the urgent need for better safeguards in medical AI. As artificial intelligence becomes more integrated into the daily routines of both doctors and patients, researchers sought to understand a critical question: would AI chatbots uncritically accept and repeat incorrect medical details embedded in a user’s query? The results were clear. 

“What we saw across the board is that AI chatbots can be easily misled by false medical details, whether those errors are intentional or accidental,” stated lead author Dr. Mahmud Omar. “They not only repeated the misinformation but often expanded on it, offering confident explanations for non-existent conditions.” This tendency for AI to “hallucinate” or generate fabricated information poses a significant risk in a medical context, where accuracy can have life-or-death consequences.

Testing the AI with “Fake” Medical Terms

To assess this vulnerability, the research team designed an experiment using fictional patient scenarios. Each scenario contained a single fabricated medical term, such as a made-up disease, symptom, or lab test. They submitted these prompts to several leading large language models (LLMs).

In the initial phase of the study, the chatbots consistently failed to recognize the fictional terms. Instead of questioning the fabricated information, they often treated it as fact and confidently generated detailed, plausible-sounding explanations about the non-existent conditions or treatments.

“Our goal was to see whether a chatbot would run with false information if it was slipped into a medical question, and the answer is yes,” said Dr. Eyal Klang, co-corresponding senior author and Chief of Generative AI in the Windreich Department of Artificial Intelligence and Human Health at Icahn Mount Sinai. “Even a single made-up term could trigger a detailed, decisive response based entirely on fiction.”

A Simple Safeguard Makes a Big Difference

In the second round of the experiment, the researchers introduced a simple but powerful change. They added a one-line caution to their prompt, reminding the AI that the information provided in the user’s query might be inaccurate.

The impact was immediate and significant. The addition of this simple warning prompt dramatically reduced the AI’s tendency to elaborate on the fake medical details.

“The encouraging part is that a simple, one-line warning added to the prompt cut those hallucinations dramatically, showing that small safeguards can make a big difference,” Dr. Omar noted. Dr. Klang added that the safety reminder “made an important difference, cutting those errors nearly in half.”

Engineering Safer AI

The Mount Sinani research team plans to continue this work by applying their “fake-term” stress-testing method to real, de-identified patient records and developing more advanced safety prompts. They hope this approach can become a standard tool for hospitals, developers, and regulators to validate AI systems before they are deployed in a clinical setting.

  • LinkedIn
  • Twitter
  • Facebook
  • Email
  • Print

Tagged With: Artificial Intelligence, Mount Sinai

Tap Native

Get in-depth healthcare technology analysis and commentary delivered straight to your email weekly

Reader Interactions

Primary Sidebar

Subscribe to HIT Consultant

Latest insightful articles delivered straight to your inbox weekly.

Submit a Tip or Pitch

Featured Interview

Reach7 Diabetes Studios Founder Chun Yong on Reimagining Chronic Care with a Concierge Medical Model

Most-Read

HHS Finalizes HTI-4 Rule: Prior Authorization & E-Prescribing Interoperability

HHS Finalizes HTI-4 Rule: Prior Authorization & E-Prescribing Interoperability

Meaningful Use Penalties_Meaningful Use_Partial Code Free_Senators Urge CMS to Establish Clear Metrics for ICD-10 Testing

CMS Finalizes TEAM Model: A New Era of Value-Based Surgical Care

White House Event Unveils CMS Health Tech Ecosystem Initiative

White House Event Unveils CMS Health Tech Ecosystem Initiative

Digital Health Faces Q2'25 Pullback: Funding Falls to 5-Year Low, But AI Dominates and $1B+ IPOs Emerge

Healthcare Investment Shifts in 1H 2025: AI Remains a Bright Spot Amidst Fundraising Decline

Digital Health Faces Q2'25 Pullback: Funding Falls to 5-Year Low

Digital Health Faces Q2’25 Pullback: Funding Falls to 5-Year Low

Beyond the Hype: Building AI Systems in Healthcare Where Hallucinations Are Not an Option

Beyond the Hype: Building AI Systems in Healthcare Where Hallucinations Are Not an Option

Health IT Sector Navigates Policy Turbulence with Resilient M&A

Health IT’s New Chapter: IPOs Return, Resilient M&A, Valuations Rise in 1H 2025

PwC Report: US Medical Cost Trend to Remain Elevated at 8.5% in 2026

PwC Report: US Medical Cost Trend to Remain Elevated at 8.5% in 2026

Philips Launches ECG AI Marketplace, Partnering with Anumana to Enhance Cardiac Care with AI-Powered Diagnostics

Philips Launches ECG AI Marketplace, Partnering with Anumana to Enhance Cardiac Care with AI-Powered Diagnostics

WeightWatchers Emerges from Bankruptcy, Launches New Menopause Program

WeightWatchers Emerges from Bankruptcy, Launches New Menopause Program

Secondary Sidebar

Footer

Company

  • About Us
  • Advertise with Us
  • Reprints and Permissions
  • Submit An Op-Ed
  • Contact
  • Subscribe

Editorial Coverage

  • Opinion
  • Health IT
    • Care Coordination
    • EMR/EHR
    • Interoperability
    • Population Health Management
    • Revenue Cycle Management
  • Digital Health
    • Artificial Intelligence
    • Blockchain Tech
    • Precision Medicine
    • Telehealth
    • Wearables
  • Startups
  • Value-Based Care
    • Accountable Care
    • Medicare Advantage

Connect

Subscribe to HIT Consultant Media

Latest insightful articles delivered straight to your inbox weekly

Copyright © 2025. HIT Consultant Media. All Rights Reserved. Privacy Policy |