ChatGPT Does a Bad Job of Answering People’s Medication Questions, Study Finds

08 Dec 2023
Researchers recently tested ChatGPT’s ability to answer patient questions about medication, finding that the viral chatbot came up dangerously short. The research was presented at the American Society of Health-System Pharmacists’ annual meeting, which was held this week in Anaheim. The free version of ChatGPT, which was the one tested in the study, has more than 100 million users. Providers should be wary of the fact that the generative AI model does not always give sound medical advice, given many of their patients could be turning to ChatGPT to answer health-related questions, the study pointed out. The study was conducted by pharmacy researchers at Long Island University. They first gathered 45 questions that patients posed to the university’s drug information service in 2022 and 2023, and then they wrote their answers to them. Each response was reviewed by a second investigator. Then, the research team fed these same questions to ChatGPT and compared the answers to the pharmacist-produced responses. The researchers gave ChatGPT 39 questions instead of 45, as the subject matter for six of the questions lacked the published literature needed for ChatGPT to provide a data-driven response. The study found that only a quarter of ChatGPT’s answers were satisfactory. ChatGPT did not directly address 11 questions, gave wrong answers to 10, and provided incomplete answers for another 12, the researchers wrote. For instance, one question asked whether there is a drug interaction between the blood-pressure lowering medication verapamil and Paxlovid, Pfizer’s antiviral pill for Covid-19. ChatGPT said that there is no interaction between the two drugs, which isn’t true — combining those two medications could dangerously lower a person’s blood pressure. In some cases, the AI model generated false scientific references to support its response. In each prompt, the researchers asked ChatGPT to show references to the information provided in its answers, but the model provided references in only eight responses — and all of those references were made-up. “Healthcare professionals and patients should be cautious about using ChatGPT as an authoritative source for medication-related information,” Dr. Sara Grossman, a lead author of the study, said in a statement. “Anyone who uses ChatGPT for medication-related information should verify the information using trusted sources.” ChatGPT’s usage policy echoes Dr. Grossman’s sentiments. It states that the model is “not fine-tuned to provide medical information,” and that people should never turn to it when seeking “diagnostic or treatment services for serious medical conditions.”
The content of the article does not represent any opinions of Synapse and its affiliated companies. If there is any copyright infringement or error, please contact us, and we will deal with it within 24 hours.
Indications
Targets
-
Chat with Hiro
Get started for free today!
Accelerate Strategic R&D decision making with Synapse, PatSnap’s AI-powered Connected Innovation Intelligence Platform Built for Life Sciences Professionals.
Start your data trial now!
Synapse data is also accessible to external entities via APIs or data packages. Empower better decisions with the latest in pharmaceutical intelligence.