Connect with us

Hi, what are you looking for?

Top Stories

AI Chatbots Produce 49.6% Problematic Health Responses in New BMJ Open Study

A new BMJ Open study reveals that five AI chatbots, including ChatGPT and Grok, deliver 49.6% problematic health responses, raising urgent oversight concerns.

Five widely used AI chatbots have been found to frequently deliver problematic answers to health-related inquiries, according to a study published in the BMJ Open on April 15, 2026. The research tested five AI models—Gemini, DeepSeek, Meta AI, ChatGPT, and Grok—with 50 prompts across five categories known for misinformation: cancer, vaccines, stem cells, nutrition, and athletic performance. The study’s findings raise significant concerns about the deployment of AI in health settings without adequate oversight.

The researchers designed the questions to challenge the chatbots with potentially misleading advice. Out of 250 total responses, nearly 50% were rated as problematic, with 30% considered somewhat problematic and 19.6% classified as highly problematic. While the analysis revealed no statistically significant differences in overall performance among the chatbots, Grok had a higher incidence of highly problematic responses.

The performance of the chatbots varied across different health categories, with stronger results observed in responses to questions about vaccines and cancer. Conversely, the chatbots struggled most with prompts related to stem cells, nutrition, and athletic performance. The study also noted that open-ended questions elicited significantly more highly problematic responses compared to closed-ended inquiries.

In terms of citation quality, the chatbots fell short. Among 25 closed-ended questions, the tools produced references roughly 81% of the time, yet the median completeness score hovered around just 40%. Notably, none of the chatbots generated a fully accurate and complete reference list, raising further concerns about the reliability of the information provided.

The readability of the responses was another issue; answers were often difficult for the average user to comprehend, requiring a higher education level for better understanding. The study’s authors expressed alarm at the implications of these findings, warning that the continued use of AI chatbots in health contexts without enhanced oversight could exacerbate the spread of misinformation.

As AI technology evolves and becomes increasingly integrated into various sectors, the implications of such findings could influence regulatory discussions around AI deployment in healthcare. Stakeholders may need to consider stringent guidelines to ensure that AI systems offer safe, accurate, and accessible information to users, particularly in sensitive areas like health and medicine. The urgency of addressing these issues highlights the need for robust oversight as reliance on AI tools continues to grow.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Finance

70% of finance teams in Australia and New Zealand use shadow AI tools like ChatGPT, risking data governance with only 16% confident in data...

AI Research

Mayo Clinic's Evo 2 AI model analyzes 128,000 genomes to identify cancer-causing mutations, revolutionizing early diagnosis and precision medicine.

Top Stories

Google's Gemini AI model claims 91% accuracy, yet it generates tens of millions of errors annually, raising alarms about misinformation in search results

Top Stories

Stanford's AI Index reveals U.S. investment of $285.9B eclipses China's $12.4B, yet 95% of AI projects see no ROI and model gap narrows to...

AI Education

Google showcases Gemini for Education and NotebookLM at key tech events, empowering students with personalized AI tools to enhance learning outcomes.

Top Stories

Therapists are urged to explore patients' AI chatbot use for emotional support, as a JAMA Psychiatry study reveals its growing role in mental health...

AI Business

Anthropic's Claude gained traction at the HumanX conference, signaling a pivotal shift in enterprise AI as businesses favor reliability over OpenAI's previous dominance.

AI Technology

Meta AI and KAUST unveil Neural Computers, a revolutionary computing model merging computation, memory, and I/O, paving the way for a new AI paradigm.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.