24 August,2023 08:24 AM IST | New York | IANS
Image for representational purposes only. Photo Courtesy: iStock
Amid the growing debate over AI and its side effects, researchers now report that in healthcare settings, ChatGPT is nearly 72 per cent accurate across all medical specialties and phases of clinical care, and 77 per cent accurate in making final diagnoses. The study led by investigators from Mass General Brigham highlighted potential for generative AI to increase access and efficiency in healthcare.
The large-language model (LLM) artificial intelligence chatbot performed equally well in both primary care and emergency settings across all medical specialties, said the study published in the Journal of Medical Internet Research.
"No real benchmarks exist, but we estimate this performance to be at the level of someone who has just graduated from medical school, such as an intern or resident. This tells us that LLMs in general have the potential to be an augmenting tool for the practice of medicine and support clinical decision-making with impressive accuracy," said Marc Succi, associate chair of innovation and commercialisation and strategic innovation leader at Mass General Brigham.
Changes in artificial intelligence technology are occurring at a fast pace and transforming many industries, including health care. But the capacity of LLMs to assist in the full scope of clinical care has not yet been studied. Succi and his team tested the hypothesis that ChatGPT would be able to work through an entire clinical encounter with a patient and recommend a diagnostic workup, decide the clinical management course, and ultimately make the final diagnosis.
ALSO READ
If you're positioned properly, it's an opportunity: Rahul shares views on AI
Apple Event 2024: New iPhone 16 set to lead AI revolution
Here's why Anil Kapoor is named on TIME's 100 most influential people in AI
New tech launch alert: Apple to launch iPhone 16 series on September 9
CARE Hospitals Bhubaneswar: Pioneering the Future of Healthcare with Artificial Intelligence and Robotics in Orthopedics
The team compared ChatGPT's accuracy on differential diagnosis, diagnostic testing, final diagnosis, and management in a structured blinded process, awarding points for correct answers and using linear regressions to assess the relationship between ChatGPT's performance and the vignette's demographic information. ChatGPT was lowest-performing in making differential diagnoses, where it was only 60 per cent accurate. It was only 68 percent accurate in clinical management decisions, such as figuring out what medications to treat the patient with after arriving at the correct diagnosis.
ChatGPT's answers did not show gender bias and that its overall performance was steady across both primary and emergency care. "ChatGPT struggled with differential diagnosis, which is the meat and potatoes of medicine when a physician has to figure out what to do," said Succi. "That is important because it tells us where physicians are truly experts and adding the most value -- in the early stages of patient care with little presenting information, when a list of possible diagnoses is needed." The authors note that before tools like ChatGPT can be considered for integration into clinical care, more benchmark research and regulatory guidance is needed.
Also Read: What exactly is ChatGPT? All you need to know about the artificial intelligence chatbot
This story has been sourced from a third party syndicated feed, agencies. Mid-day accepts no responsibility or liability for its dependability, trustworthiness, reliability and data of the text. Mid-day management/mid-day.com reserves the sole right to alter, delete or remove (without notice) the content in its absolute discretion for any reason whatsoever