Digital Healthcare Insights: January 22 - January 28
GPT-4o now outperforms human radiologists at complex CT protocoling
A new study suggests that artificial intelligence may soon
handle the complex administrative tasks that currently bog down radiologists.
Researchers tested GPT-4o on its ability to assign the correct scanning
protocols for abdominal and pelvic CTs—a critical step that determines how the
machine is set up for each patient. The AI achieved an impressive 96.2%
accuracy rate significantly outperforming the human radiologists who scored
88.3%. The model was particularly effective when provided with specific clinical
context demonstrating that large language models are now capable of
understanding nuanced medical instructions. This finding points to a future
where AI handles the workflow logistics allowing human doctors to focus
entirely on image interpretation and diagnosis.
Read the original article at: https://radiologybusiness.com/topics/artificial-intelligence/gpt-4o-outperforms-radiologists-ct-protocoling
A new study shows the open-source model DeepSeek-R1 beating GPT-4o in
answering complex heart disease questions
In a surprising upset for proprietary tech giants an open
source AI model has taken the lead in patient communication. A 2026 study
evaluated how well various AI models could answer patient inquiries about
atherosclerotic cardiovascular disease (ASCVD). The open source model DeepSeek
R1 achieved a 96% "good response" rate outperforming both OpenAI's
GPT-4o and Google's Gemini. The study highlights that specialized open source
tools can be just as effective as expensive commercial models for patient
education. However researchers noted a critical flaw across the board: all
models, including DeepSeek, struggled to accurately provide specific treatment
regimens reinforcing that while AI is an excellent educator it is not yet ready
to be a prescriber.
Read the original article at: https://medinform.jmir.org/2026/1/e81422
New trials compare Claude, ChatGPT, and DeepSeek on mammography reports
While AI is excelling at administrative tasks it still
struggles to beat the human eye in complex cancer diagnosis. A comparative
study pitted three major language models—ChatGPT-4o, Claude 3 Opus, and
DeepSeek-R1—against human radiologists in analyzing mammography reports for
breast cancer risks (BI-RADS 4). The results were clear: human radiologists
significantly outperformed all three AI models. While the AI tools demonstrated
high sensitivity meaning they were good at flagging potential issues they suffered
from low specificity leading to a high volume of false alarms. The authors
conclude that for now these tools are best used as "safety net"
assistants to ensure nothing is missed rather than as independent diagnostic
agents.
Read the original article at: https://medinform.jmir.org/2025/1/e80182
Misuse of AI Chatbots has been officially named the top Health Tech Hazard
of 2026
The ECRI Institute, a global authority on medical safety,
has designated the misuse of AI chatbots as the number one health technology
hazard for 2026. The report warns that the rapid deployment of patient-facing
AI tools is creating a dangerous "credibility trap." These chatbots
often produce "coherent nonsense"—answers that sound authoritative
and grammatically perfect but are factually incorrect. Because the output looks
professional patients are less likely to question it leading to potential
medication errors or delayed care. The institute urges healthcare organizations
to implement strict oversight and visible disclaimers before letting these
tools interact directly with patients.
Read the original article at: https://www.healthcareinfosecurity.com/healthcare-chatbots-provoke-unease-in-ai-governance-analysts-a-30483
Follow us on
Instagram, Twitter, and Facebook to stay up to date
with what's new in healthcare all around the world.
Comments
Post a Comment