Applicability of Online Chat-Based Artificial Intelligence Models to Colorectal Cancer Screening.

MedStar author(s):
Citation: Digestive Diseases & Sciences. 69(3):791-797, 2024 Mar.PMID: 38267726Institution: MedStar Union Memorial HospitalDepartment: Gastroenterology Fellowship | Hospitalist | Internal Medicine Residency | Medicine | MedStar Georgetown University Hospital/MedStar Washington Hospital CenterForm of publication: Journal ArticleMedline article type(s): Journal ArticleSubject headings: *Colorectal Neoplasms | *Early Detection of Cancer | Artificial Intelligence | Colorectal Neoplasms/di [Diagnosis] | Communication | Humans | Reproducibility of Results | Year: 2024Local holdings: Available online from MWHC library: 1996 - present, Available in print through MWHC library: 1999 - 2006ISSN:
  • 0163-2116
Name of journal: Digestive diseases and sciencesAbstract: BACKGROUND: Over the past year, studies have shown potential in the applicability of ChatGPT in various medical specialties including cardiology and oncology. However, the application of ChatGPT and other online chat-based AI models to patient education and patient-physician communication on colorectal cancer screening has not been critically evaluated which is what we aimed to do in this study.CONCLUSION: Most responses provided by AI models on CRC screening were appropriate. Some limitations exist in their ability to correctly interpret medical literature and provide updated information in answering queries. Patients should consult their physicians for context on the recommendations made by these AI models. Copyright © 2024. The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.METHODS: We posed 15 questions on important colorectal cancer screening concepts and 5 common questions asked by patients to the 3 most commonly used freely available artificial intelligence (AI) models. The responses provided by the AI models were graded for appropriateness and reliability using American College of Gastroenterology guidelines. The responses to each question provided by an AI model were graded as reliably appropriate (RA), reliably inappropriate (RI) and unreliable. Grader assessments were validated by the joint probability of agreement for two raters.RESULTS: ChatGPT and YouChat TM provided RA responses to the questions posed more often than BingChat. There were two questions that > 1 AI model provided unreliable responses to. ChatGPT did not provide references. BingChat misinterpreted some of the information it referenced. The age of CRC screening provided by YouChat TM was not consistently up-to-date. Inter-rater reliability for 2 raters was 89.2%.All authors: Atarere J, Naqvi H, Haas C, Adewunmi C, Bandaru S, Allamneni R, Ugonabo O, Egbo O, Umoren M, Kanth PFiscal year: FY2024Digital Object Identifier: ORCID: Date added to catalog: 2024-04-24
Holdings
Item type Current library Collection Call number Status Date due Barcode
Journal Article MedStar Authors Catalog Article 38267726 Available 38267726

Available online from MWHC library: 1996 - present, Available in print through MWHC library: 1999 - 2006

BACKGROUND: Over the past year, studies have shown potential in the applicability of ChatGPT in various medical specialties including cardiology and oncology. However, the application of ChatGPT and other online chat-based AI models to patient education and patient-physician communication on colorectal cancer screening has not been critically evaluated which is what we aimed to do in this study.

CONCLUSION: Most responses provided by AI models on CRC screening were appropriate. Some limitations exist in their ability to correctly interpret medical literature and provide updated information in answering queries. Patients should consult their physicians for context on the recommendations made by these AI models. Copyright © 2024. The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.

METHODS: We posed 15 questions on important colorectal cancer screening concepts and 5 common questions asked by patients to the 3 most commonly used freely available artificial intelligence (AI) models. The responses provided by the AI models were graded for appropriateness and reliability using American College of Gastroenterology guidelines. The responses to each question provided by an AI model were graded as reliably appropriate (RA), reliably inappropriate (RI) and unreliable. Grader assessments were validated by the joint probability of agreement for two raters.

RESULTS: ChatGPT and YouChat TM provided RA responses to the questions posed more often than BingChat. There were two questions that > 1 AI model provided unreliable responses to. ChatGPT did not provide references. BingChat misinterpreted some of the information it referenced. The age of CRC screening provided by YouChat TM was not consistently up-to-date. Inter-rater reliability for 2 raters was 89.2%.

English

Powered by Koha