Oops, you're using an old version of your browser so some of the features on this page may not be displaying properly.

MINIMAL Requirements: Google Chrome 24+Mozilla Firefox 20+Internet Explorer 11Opera 15–18Apple Safari 7SeaMonkey 2.15-2.23

Poster session 07

2162P - Can artificial intelligence provide accurate and reliable answers to cancer patients' questions? Comparison of chatbots based on the ESMO Patient Guide about cancer pain

Date

21 Oct 2023

Session

Poster session 07

Topics

Supportive Care and Symptom Management

Tumour Site

Presenters

Kadriye Bir Yucel

Citation

Annals of Oncology (2023) 34 (suppl_2): S1080-S1134. 10.1016/S0923-7534(23)01268-1

Authors

K. Bir Yucel1, O. Yazici1, N. Ozdemir2, A. Özet1

Author affiliations

  • 1 Medical Oncology Department, Gazi University - Faculty of Medicine, 06560 - Ankara/TR
  • 2 Medical Oncology Department, Ankara Numune Education and Research Hospital, 06100 - Ankara/TR

Resources

Login to get immediate access to this content.

If you do not have an ESMO account, please create one for free.

Abstract 2162P

Background

There is limited data on the quality of cancer information provided by ChatGPT and other artificial intelligence systems. In this study, we aimed to compare the accuracy of information about cancer pain provided by chatbots (chatGPT, perplexity, and chatsonic) based on the questions and answers contained in the the European Medical Oncology Association (ESMO) Patient Guide about cancer pain.

Methods

Twenty questions were selected from the questions available in the ESMO Patient Guide Cancer Pain. Medical oncologists with more than 10 years of experience compared responses from chatbots (ChatGPT, Perplexity, and Chatsonic) with the ESMO patient guide. The primary evaluation criteria for the quality of the responses were accuracy, patient readability, and stability of response. The accuracy of responses was evaluated using a three-point scale: 1 for accuracy, 2 for a mixture of accurate and incorrect or outdated data, and 3 for wholly inaccurate. The Flesch-Kincaid readability (FKr) grade was used to measure readability. Stability of responses was evaluated whether the model’s accuracy is consistent across different answers to the same question.

Results

Chatbots were more difficult to read than the ESMO patient guideline (FKr= 9.6 vs. 12.8, p= 0.072). Among the chatbots, perplexity had the easiest readability (FKr= 11.2). In the accuracy evaluation, the percentage of overall agreement for accuracy was 100% for ESMO answers and 96% for ChatGPT outputs for questions (k= 0.03, standard error= 0.08). Among the chatbots, the most accurate information was obtained with chatGPT.

Table: 2162P

Comparison of ESMO and chatbots in terms of readibility and accuracy

ESMO Chatbots
ChatGPT Perplexity Chatsonic p-value
Readibility (FKr grade) 9.6 Easily understood 13.4 Difficult to read 11.2 Fairly difficult to read 13.9 Difficult to read 0.072
Accuracy %100 %96 %86 % 90 0.037

Conclusions

The results suggest that ChatGPT provides more accurate information about cancer pain compared with other chatbots.

Clinical trial identification

Editorial acknowledgement

Legal entity responsible for the study

Gazi University Ethic Commitee.

Funding

Has not received any funding.

Disclosure

All authors have declared no conflicts of interest.

This site uses cookies. Some of these cookies are essential, while others help us improve your experience by providing insights into how the site is being used.

For more detailed information on the cookies we use, please check our Privacy Policy.

Customise settings
  • Necessary cookies enable core functionality. The website cannot function properly without these cookies, and you can only disable them by changing your browser preferences.