Full Text

Turn on search term navigation

© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

Simple Summary

Cancer patients, especially long-distance patients, often struggle to receive timely and precise medical information and support for their symptom management and survivorship care. ChatGPT-4’s responses to queries concerning head and neck (HN) cancer remain questionable. The purpose of this study was to develop and validate a retrained large language model (LLM) for HN cancer patients. In this cross-sectional study, the presented LLM was retrained with a high-quality user-defined knowledge base. The responses from the LLM to patients’ questions were validated against human responses, and the model showed a superior performance, with average scores of 4.25 for accuracy, 4.35 for clarity, 4.22 for completeness, and 4.32 for relevance, on a 5-point scale. The confined-trained LLM with a high-quality user-defined knowledge base demonstrates high accuracy, clarity, completeness, and relevance in offering evidence-based information and guidance on the symptom management and survivorship care for head and neck cancer patients.

Abstract

Purpose: This study aimed to develop a retrained large language model (LLM) tailored to the needs of HN cancer patients treated with radiotherapy, with emphasis on symptom management and survivorship care. Methods: A comprehensive external database was curated for training ChatGPT-4, integrating expert-identified consensus guidelines on supportive care for HN patients and correspondences from physicians and nurses within our institution’s electronic medical records for 90 HN patients. The performance of our model was evaluated using 20 patient post-treatment inquiries that were then assessed by three Board certified radiation oncologists (RadOncs). The rating of the model was assessed on a scale of 1 (strongly disagree) to 5 (strongly agree) based on accuracy, clarity of response, completeness s, and relevance. Results: The average scores for the 20 tested questions were 4.25 for accuracy, 4.35 for clarity, 4.22 for completeness, and 4.32 for relevance, on a 5-point scale. Overall, 91.67% (220 out of 240) of assessments received scores of 3 or higher, and 83.33% (200 out of 240) received scores of 4 or higher. Conclusion: The custom-trained model demonstrates high accuracy in providing support to HN patients offering evidence-based information and guidance on their symptom management and survivorship care.

Details

Title
Testing and Validation of a Custom Retrained Large Language Model for the Supportive Care of HN Patients with External Knowledge Base
Author
Zhu, Libing  VIAFID ORCID Logo  ; Yi, Rong; McGee, Lisa A; Rwigema, Jean-Claude M; Patel, Samir H
First page
2311
Publication year
2024
Publication date
2024
Publisher
MDPI AG
e-ISSN
20726694
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
3078990927
Copyright
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.