AI medical advice: Not ready for prime time

Credit: Stanford
Credit: Stanford

Medical artificial intelligence (AI) is often described as a way to make patient care safer by helping clinicians manage information. A new study by the Icahn School of Medicine at Mount Sinai and collaborators confronts a critical vulnerability: when a medical lie enters the system, can AI pass it on as if it were true?  

Analyzing more than a million prompts across nine leading language models, the researchers found that these systems can repeat false medical claims when they appear in realistic hospital notes or social-media health discussions. 

The findings, published in the February 9 online issue of The Lancet Digital Health [10.1016/j.landig.2025.100949], suggest that current safeguards do not reliably distinguish fact from fabrication once a claim is wrapped in familiar clinical or social-media language. 

To test this systematically, the team exposed the models to three types of content: real hospital discharge summaries from the Medical Information Mart for Intensive Care (MIMIC) database with a single fabricated recommendation added; common health myths collected from Reddit; and 300 short clinical scenarios written and validated by physicians. Each case was presented in multiple versions, from neutral wording to emotionally charged or leading phrasing similar to what circulates on social platforms. 

In one example, a discharge note falsely advised patients with esophagitis-related bleeding to “drink cold milk to soothe the symptoms.” Several models accepted the statement rather than flagging it as unsafe. They treated it like ordinary medical guidance. 

“Our findings show that current AI systems can treat confident medical language as true by default, even when it’s clearly wrong,” says co-senior and co-corresponding author Eyal Klang, MD, Chief of Generative AI in the Windreich Department of Artificial Intelligence and Human Health at the Icahn School of Medicine at Mount Sinai. “A fabricated recommendation in a discharge note can slip through. It can be repeated as if it were standard care. For these models, what matters is less whether a claim is correct than how it is written.”  

Follow the latest news and policy debates on sustainable agriculture, biomedicine, and other ‘disruptive’ innovations. Subscribe to our newsletter.

The authors say the next step is to treat “can this system pass on a lie?” as a measurable property, using large-scale stress tests and external evidence checks before AI is built into clinical tools. 

“Hospitals and developers can use our dataset as a stress test for medical AI,” says physician-scientist and first author Mahmud Omar, MD, who consults with the research team. “Instead of assuming a model is safe, you can measure how often it passes on a lie, and whether that number falls in the next generation.”  

“AI has the potential to be a real help for clinicians and patients, offering faster insights and support,” says co-senior and co-corresponding author Girish N. Nadkarni, MD, MPH, Chair of the Windreich Department of Artificial Intelligence and Human Health, Director of the Hasso Plattner Institute for Digital Health, Irene and Dr. Arthur M. Fishberg Professor of Medicine at the Icahn School of Medicine at Mount Sinai, and Chief AI Officer of the Mount Sinai Health System. “But it needs built-in safeguards that check medical claims before they are presented as fact. Our study shows where these systems can still pass on false information, and points to ways we can strengthen them before they are embedded in care.” 

The paper is titled “Mapping LLM Susceptibility to Medical Misinformation Across Clinical Notes and Social Media.”  

A version of this article was originally posted at Mount Sinai and is reposted here. Any reposting should credit both the GLP and original article. Find Mount Sinai on X @MountSinaiNYC

{{ reviewsTotal }}{{ options.labels.singularReviewCountLabel }}
{{ reviewsTotal }}{{ options.labels.pluralReviewCountLabel }}
{{ options.labels.newReviewButton }}
{{ userData.canReview.message }}

Related Articles

Infographic: Global regulatory and health research agencies on whether glyphosate causes cancer

Infographic: Global regulatory and health research agencies on whether glyphosate causes cancer

Does glyphosate—the world's most heavily-used herbicide—pose serious harm to humans? Is it carcinogenic? Those issues are of both legal and ...

Most Popular

Screenshot-2026-05-01-at-11.56.24-AM
‘Science moves forward when people are willing to think differently’: Memories of DNA maverick Craig Venter
Screenshot-2026-04-03-at-11.15.51-AM
Paraben panic: How a flawed study, media hype, and chemophobia convinced the public of the danger of one of the safest classes of preservatives
79d03212-2508-45d0-b427-8e9743ff6432
Viewpoint: The Casey Means hustle—Wellness woo opportunism dressed up as medical wisdom
Screenshot-PM-24
Viewpoint: The herbicide glyphosate isn’t perfect. Banning it would be far worse.
ChatGPT-Image-Apr-30-2026-12_21_05-PM-2
The tech billionaires behind the immortality movement
ChatGPT-Image-Mar-27-2026-11_27_05-AM
The myths of “process”: What science says about the “dangers’ of synthetic products and ultra-processed foods
ChatGPT-Image-Mar-10-2026-01_39_01-PM
Viewpoint—“Miracle molecule” debunked: Why acemannan supplements don’t work
Drinking lots of water can help reduce the effects of aging
Nanoplastics in drinking water: MAHA activists forge science-based bipartisan coalition 
ChatGPT-Image-May-1-2026-03_16_32-PM
Viewpoint: How ‘health care guru’ Joe Rogan circumvented the FDA’s skepticism on psychedelics
ChatGPT Image Apr 30, 2026, 01_11_41 PM
Trump’s America First health aid cuts: Retrenchment has already hit global malaria, HIV, TB, and polio programs
Screenshot-2026-04-12-135256
Bixonimania: The fake disease scam that AI swallowed whole
ChatGPT-Image-Mar-3-2026-01_17_14-PM
MAHA wellness influencers deride proven anxiety medications, tout lifestyle fixes
images
The never-ending GMO debate: Pros and cons
glp menu logo outlined

Get news on human & agricultural genetics and biotechnology delivered to your inbox.