Breaking research

ChatGPT is Debunking Myths on Social Media Around Vaccine Safety, Say Experts

Peer-reviewed Survey

This additional information is included as part of a press release labeling system introduced by the Academy of Medical Sciences and Science Media Centre. For more details click here.

Open Access Logo

Photo by Matheus Bertelli on Pexels

ChatGPT could help to increase vaccine uptake by debunking myths around jab safety, say the authors of a study published in the peer-reviewed journal Human Vaccines and Immunotherapeutics.

The researchers asked the artificial intelligence (AI) chatbot the top 50 most frequently-asked Covid-19 vaccine questions. They included queries based on myths and fake stories such as the vaccine causing Long Covid.

Results show that ChatGPT scored nine out of 10 on average for accuracy. The rest of the time it was correct but left some gaps in the information provided, according to the study.

Based on these findings, experts who led the study from the GenPoB research group based at the Instituto de Investigación Sanitaria (IDIS) – Hospital Clinico Universitario of Santiago de Compostela, say the AI tool is a “reliable source of non-technical information to the public”, especially for people without specialist scientific knowledge.

However, the findings do highlight some concerns about the technology such as ChatGPT changing its answers in certain situations.

“Overall, ChatGPT constructs a narrative in line with the available scientific evidence, debunking myths circulating on social media,” says lead author Antonio Salas, who as well as leading the GenPoB research group, is also a Professor at the Faculty of Medicine at the University of Santiago de Compostela, in Spain.

“Thereby it potentially facilitates an increase in vaccine uptake. ChatGPT can detect counterfeit questions related to vaccines and vaccination. The language this AI uses is not too technical and therefore easily understandable to the public but without losing scientific rigor.

“We acknowledge that the present-day version of ChatGPT cannot substitute an expert or scientific evidence. But the results suggest it could be a reliable source of information to the public.”

In 2019, the World Health Organisation (WHO) listed vaccine hesitancy among the top 10 threats to global health.

During the pandemic, misinformation spread via social media contributed to public mistrust of Covid-19 vaccination.

The authors of this study include those from the Hospital Clinico Universitario de Santiago which the WHO designated as a vaccine safety collaborating center in 2018.

Researchers at the center have been exploring myths around vaccine safety and medical situations that are falsely believed to be a reason not to get vaccinated. These misplaced concerns contribute to vaccine hesitancy.

The study authors set out to test ChatGPT’s ability to get the facts right and share accurate information around Covid vaccine safety in line with current scientific evidence.

ChatGPT enables people to have human-like conversations and interactions with a virtual assistant. The technology is very user-friendly which makes it accessible to a wide population.

However, many governments are concerned about the potential for ChatGPT to be used fraudulently in educational settings such as universities.

The study was designed to challenge the chatbot by asking it the questions most frequently received by the WHO collaborating center in Santiago.

The queries covered three themes. The first was misconceptions around safety such as the vaccine causing Long Covid. Next was false contraindications – medical situations where the jab is safe to use such as in breastfeeding women.

The questions also related to true contraindications – a health condition where the vaccine should not be used – and cases where doctors must take precautions eg a patient with heart muscle inflammation.

Next, experts analyzed the responses then rated them for veracity and precision against current scientific evidence, and recommendations from WHO and other international agencies.

The authors say this was important because algorithms created by social media and internet search engines are often based on an individual’s usual preferences. This may lead to ‘biased or wrong answers’, they add.

Results showed that most of the questions were answered correctly with an average score of nine out of 10 which is defined as ‘excellent’ or ‘good’. The responses to the three question themes were on average 85.5% accurate or 14.5% accurate but with gaps in the information provided by ChatGPT.

ChatGPT provided correct answers to queries that arose from genuine vaccine myths, and to those considered in clinical recommendation guidelines to be false or true contraindications.

However, the research team does highlight ChatGPT’s downsides in providing vaccine information.

Professor Salas, who specializes in human genetics, concludes: “Chat GPT provides different answers if the question is repeated ‘with a few seconds of delay’.

“Another concern we have seen is that this AI tool, in its present version, could also be trained to provide answers not in line with scientific evidence.

“One can ‘torture’ the system in such a way that it will provide the desired answer. This is also true for other contexts different to vaccines. For instance, it might be possible to make the chatbot align with absurd narratives like the flat-earth theory, deny climate change, or object to the theory of evolution, just to give a few examples.

“However, it’s important to note that these responses are not the default behaviour of ChatGPT. Thus, the results we have obtained regarding vaccine safety can be probably extrapolated to many other myths and pseudoscience.”