ChatGPT-4 matches radiologists in flagging errors on reports

In this study, the researchers intentionally inserted 150 errors from five error categories (omission, insertion, spelling, side confusion and “other”) into 100 of the 200 reports and tasked the ChatGPT-4 and two senior radiologists, two attending physicians, and two residents with detecting these errors. ChatGPT-4’s detection rate was 82.7% (124 of 150), while the error detection rates were 89.3% for senior radiologists (134 out of 150) and 80% for attending radiologists and radiology residents (120 out of 150), on average, the researchers found. In addition, GPT-4 required less processing time per radiology report than even the fastest human reader, and the use of GPT-4 resulted in lower mean correction cost per report than the most cost-efficient radiologist, Gertz and Kottlors noted. The group has been exploring the use of ChatGPT in radiology applications for more than a year, and is “still shocked” by its performance, given that the LLM’s developer OpenAI.com has kept a lid on the data it used to train the model, Gertz said. The study also suggests that ChatGPT-4 could potentially serve as a teaching tool for residents who might not have access to senior radiologists by providing a “feedback loop” for them to learn from their mistakes, Gertz said. Could ChatGPT-4 be used as a training tool to improve reporting accuracy? Ultimately, the study shows that the advanced text-processing capabilities of LLMs such as GPT-4 have the potential to enhance...
Source: AuntMinnie.com Headlines - Category: Radiology Authors: Tags: Artificial Intelligence Source Type: news