24 x 7 World News

Is ChatGPT in Your DoctorтАЩs Inbox?

0

May 3, 2023 — What happens when a chatbot slips into your doctorтАЩs direct messages? Depending on who you ask, it might improve outcomes. On the other hand, it might raise a few red flags.

The fallout from the COVID-19 pandemic has been far-reaching, especially when it comes to the frustration over the inability to reach a doctor for an appointment, let alone get answers to health questions. And with the rise of telehealth and a substantial increase in electronic patient messages over the past 3 years, inboxes are filling fast at the same time that doctor burnout is on the rise.

The old adage that timing is everything applies, especially since technological advances in artificial intelligence, or AI, have been rapidly gaining speed over the past year. The solution to overfilled inboxes and delayed responses may lie with the AI-powered┬аChatGPT, which was shown to substantially improve the quality and tone of responses to patient questions, according to study findings published in┬аJAMA Internal Medicine.┬а

тАЬThere are millions of people out there who canтАЩt get answers to the questions that they have, and so they post them on public social media forums like Reddit Ask Docs and hope that sometime, somewhere, an anonymous doctor will reply and give them the advice that they are looking for,тАЭ said John Ayers, PhD, lead study author and computational epidemiologist at the Qualcomm Institute at the University of California-San Diego.

тАЬAI-assisted messaging means that doctors spend less time worried about verb conjugation and more time worried about medicine,тАЭ he said.┬а

r/Askdocs vs. Ask Your Doctor

Ayers is referring to the Reddit subforum┬аr/Askdocs, a platform devoted to providing patients with answers to their most pressing medical and health questions with guaranteed anonymity. The forum has 450,000 members, and at least 1,500 are actively online at any given time.

For the study, he and his colleagues randomly selected 195 Reddit exchanges (consisting of unique patient questions and doctor answers) from last OctoberтАЩs forums, and then fed each full text question into a fresh chatbot session (meaning that it was free of any prior questions that could bias the results). The question, doctor response, and chatbot response were then stripped of any information that might indicate who (or what) was answering the question тАУ and subsequently reviewed by a team of three licensed health care professionals.┬а

тАЬOur early study shows surprising results,тАЭ said Ayers, pointing to findings that showed that health care professionals overwhelmingly preferred chatbot-generated responses over the physician responses 4 to 1.┬а

The reasons for the preference were simple: better quantity, quality, and empathy. Not only were the chatbot responses significantly longer (mean 211 words to 52 words) than doctors, ┬аbut the proportion of doctor responses that were considered “less than acceptable” in quality was over 10-fold higher than the chatbot (which were mostly “better than good”). And compared to doctors’ answers, chatbot responses were more often rated significantly higher in terms of bedside manner, resulting in a 9.8-fold greater prevalence of “empathetic” or “very empathetic” ratings.

A World of Possibilities

The past decade has demonstrated that there is a world of possibilities for AI applications, from creating mundane virtual taskmasters (like AppleтАЩs Siri or AmazonтАЩs Alexa) to┬аredressing inaccuracies in histories of past civilizations.

In health care, AI/machine learning models are being integrated into diagnosis and data analysis, e.g., to speed up X-ray, computed tomography, and magnetic resonance imaging analysis or help researchers and clinicians collate and sift through┬аreams of genetic and other types of data to learn more about the connections between diseases and fuel discovery.

тАЬThe reason why this is a timely issue now is that the release of ChatGPT has made AI finally accessible for millions of physicians,тАЭ said Bertalan Mesk├│ MD, PhD, director of The Medical Futurist Institute. тАЬWhat we need now is not better technologies, but preparing the health care workforce for using such technologies.тАЭ

Mesk├│ believes that an important role for AI lies in automating data-based or repetitive tasks, noting тАЬany technology that improves the doctor-patient relationship has a place in health care,тАЭ also highlighting the need for тАЬAI- based solutions that improve their relationship by giving them more time and attention to dedicate to each other.тАЭ

The “how” of integration will be key.

тАЬI think that there are definitely opportunities for AI to mitigate issues around physician burnout and give them more time with their patients,тАЭ said Kelly Michelson, MD, MPH, director of the Center for Bioethics and Medical Humanities at Northwestern University Feinberg School of Medicine and attending physician at Ann & Robert H. Lurie ChildrenтАЩs Hospital of Chicago. тАЬBut thereтАЩs a lot of subtle nuances that clinicians consider when theyтАЩre interacting with patients that, at least right now, don’t seem to be things that can be translated through algorithms and AI.тАЭ

If anything, Michelson said that she would argue that at this stage, AI needs to be an adjunct.

тАЬWe need to think carefully about how we incorporate it and not just use it to take over one thing until itтАЩs been better tested, including message response,тАЭ she said.┬а

┬аAyers agreed.┬а

тАЬItтАЩs really just a phase zero study. And it shows that we should now move toward patient-centered studies using these technologies and not just willy-nilly flip the switch.тАЭ

The Patient Paradigm

When it comes to the patient side of ChatGPT messaging, several questions come to mind, including relationships with their health care providers.

тАЬPatients want the ease of Google but the confidence that only their own provider may provide in answering,тАЭ said Annette Ticoras, MD, a board-certified patient advocate serving the greater Columbus, OH, area.┬а

тАЬThe goal is to ensure that clinicians and patients are exchanging the highest quality information.The messages to patients are only as good as the data that was utilized to give a response,тАЭ she said.┬а

This is especially true with regard to bias.

тАЬAI tends to be kind of generated by existing data, and so if there are biases in existing data, those biases get perpetuated in the output developed by AI,тАЭ said Michelson, referring to a concept called тАЬthe black box.тАЭ┬а

тАЬThe thing about the more complex AI is that oftentimes we canтАЩt discern whatтАЩs driving it to make a particular decision,” she said. ┬атАЬYou canтАЩt always figure out whether or not that decision is based on existing inequities in the data or some other underlying issue.тАЭ

Still, Michelson is hopeful.┬а┬а

тАЬWe need to be huge patient advocates and make sure that whenever and however AI is incorporated into health care, that we do it in a thoughtful, evidence-based way that doesnтАЩt take away from the essential human component that exists in medicine,” she said.┬а

Leave a Reply