Four recent studies highlight significant concerns regarding the reliability of large language models for medical advice, with nearly half of responses from popular chatbots like Gemini, ChatGPT, and Meta AI being problematic. These models often exhibit overconfidence, hallucinations, and fabricated citations, leading to potential misinformation amplification. Research indicates that current LLMs are not yet suitable for unsupervised patient-facing clinical decision-making, as they struggle with diagnostic reasoning and can misidentify serious conditions, raising safety concerns for widespread deployment. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Confirms that current LLMs are not safe for unsupervised patient-facing medical advice, highlighting risks of misinformation and undertriage.
RANK_REASON Multiple studies published in peer-reviewed medical journals evaluate the accuracy and safety of LLMs for medical advice.