It appears you don't have support to open PDFs in this web browser. To view this file, Open with your PDF reader
Abstract
Artificial intelligence (AI) algorithms evaluating [supine] chest radiographs ([S]CXRs) have remarkably increased in number recently. Since training and validation are often performed on subsets of the same overall dataset, external validation is mandatory to reproduce results and reveal potential training errors. We applied a multicohort benchmarking to the publicly accessible (S)CXR analyzing AI algorithm CheXNet, comprising three clinically relevant study cohorts which differ in patient positioning ([S]CXRs), the applied reference standards (CT-/[S]CXR-based) and the possibility to also compare algorithm classification with different medical experts’ reading performance. The study cohorts include [1] a cohort, characterized by 563 CXRs acquired in the emergency unit that were evaluated by 9 readers (radiologists and non-radiologists) in terms of 4 common pathologies, [2] a collection of 6,248 SCXRs annotated by radiologists in terms of pneumothorax presence, its size and presence of inserted thoracic tube material which allowed for subgroup and confounding bias analysis and [3] a cohort consisting of 166 patients with SCXRs that were evaluated by radiologists for underlying causes of basal lung opacities, all of those cases having been correlated to a timely acquired computed tomography scan (SCXR and CT within < 90 min). CheXNet non-significantly exceeded the radiology resident (RR) consensus in the detection of suspicious lung nodules (cohort [1], AUC AI/RR: 0.851/0.839, p = 0.793) and the radiological readers in the detection of basal pneumonia (cohort [3], AUC AI/reader consensus: 0.825/0.782, p = 0.390) and basal pleural effusion (cohort [3], AUC AI/reader consensus: 0.762/0.710, p = 0.336) in SCXR, partly with AUC values higher than originally published (“Nodule”: 0.780, “Infiltration”: 0.735, “Effusion”: 0.864). The classifier “Infiltration” turned out to be very dependent on patient positioning (best in CXR, worst in SCXR). The pneumothorax SCXR cohort [2] revealed poor algorithm performance in CXRs without inserted thoracic material and in the detection of small pneumothoraces, which can be explained by a known systematic confounding error in the algorithm training process. The benefit of clinically relevant external validation is demonstrated by the differences in algorithm performance as compared to the original publication. Our multi-cohort benchmarking finally enables the consideration of confounders, different reference standards and patient positioning as well as the AI performance comparison with differentially qualified medical readers.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 University Hospital, LMU Munich, Department of Radiology, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X)
2 University Hospital, LMU Munich, Department of Radiology, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X); German Center for Lung Research, Comprehensive Pneumology Center, Munich, Germany (GRID:grid.452624.3)
3 University Hospital, LMU Munich, Department of Medicine I, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X)
4 University Hospital, LMU Munich, Department of Medicine II, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X)
5 Musculoskeletal University Center Munich (MUM), University Hospital, LMU Munich, Department of Orthopaedics and Trauma Surgery, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X)
6 University Hospital, LMU Munich, Department of Radiology, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X); German Center for Lung Research, Comprehensive Pneumology Center, Munich, Germany (GRID:grid.452624.3); Asklepios Fachklinik München, Department of Radiology, Gauting, Germany (GRID:grid.452624.3)
7 University Hospital, LMU Munich, Department of Radiology, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X); University Hospital, LMU Munich, Institute of Neuroradiology, Munich, Germany (GRID:grid.5252.0) (ISNI:0000 0004 1936 973X)