1. Introduction
Chest radiography is one of the most widely used medical imaging modalities in the United States [1] as it is affordable and quick, and can rule out several critical pathological conditions such as COVID-19 pneumonia [2–5], lung cancer [6–8], pulmonary tuberculosis [9], and heart diseases [10]. In South Korea, the number of chest radiograph prescriptions has steadily increased from 23 million per year in 2015 to almost 27 million in 2019 [11].
The clinical benefit of chest radiography is user dependent [12]. Chest radiography holds a vast amount of information in a single image, which requires extensive medical training for accurate interpretation. Training a radiologist requires years, and the number of experts who can read chest radiographs is not increasing as rapidly as the number of prescriptions. In many economies, this gap in the supply and demand of medical professionals is worse in medically underserved rural areas [12–14]. Due to urban migration, these areas often have elderly populations, who are at an elevated risk of cardiovascular and pulmonary maladies that can be screened through chest radiographs, However, due to an imbalance in the distribution of healthcare resources, such clinical demands are rarely met. Medical artificial intelligence (AI) can address such shortages of clinical expertise if it can be trained to achieve levels of performance comparable to experienced physicians in interpreting specific diagnostic modalities [1]. With appropriate development and distribution of medical AI as computer-aided diagnosis programs, the unmet needs for clinical expertise in medically underserved areas can be better addressed [15].
With the rapid advancement of artificial intelligence, computer-aided diagnosis research has recently bloomed [1, 11]. The same is true with the development of deep learning, which has resulted in remarkable progress being made in the field of medical image analysis [16]. One of the predominant challenges to AI training in the medical imaging is data imbalance. To ensure that the AI model is robust and effective, the training dataset must be extensive and well balanced [17]; however, the amount of medical data is usually insufficient and the imbalance, large [18]. In many cases, the most pernicious diseases have the lowest prevalence, leading to severe data imbalances [19], which can negatively affect model performances.
Many chest radiograph datasets are publicly available, and most are prone to data imbalance. Chest X-ray 8 (CXR8), one of the largest existing open-source chest radiograph data initiatives led by the National Institute of Health, is no exception; of the 108,948 chest anteroposterior radiographs in the dataset, 84,312 (77.39%) images had no pathologic lesions, while only 1,062 (0.97%) images showed signs of pneumonia [20]. MIMIC-CXR, another extensive chest radiograph dataset created by the MIT Laboratory for Computational Physiology and Beth Israel Deaconess Medical Center, comprises 377,110 images. Of these images, only 10,801 (2.86%) contained pulmonary masses and 25,038 (6.64%) contained evidence of pneumonia [21].
With the evolution of medical AI research, many data augmentation methods have been developed to overcome such data imbalances [18]. In computer vision, common traditional data augmentation methods include flipping, rotating, shifting, and color-transforming images to oversample underrepresented data [17, 22], while in some medical images, including chest radiographs, traditional augmentation has been proven to result in worse model performance [11]. To overcome this, many novel data augmentation techniques have been developed including generative adversarial networks (GAN).
GAN is a powerful method to generate novel images without supervision [16, 23]. The algorithm employs two competing ‘adversarial’ networks: generator network G(z) and discriminator network D(x) [24]. G(z) produces realistic images and attempts to deceive D(x), while D(x) learns to better discriminate between real and fake images [25]. While G(z) seeks to minimize the cost value function V(D, G), D(x) seeks to maximize it. Consequently, GAN learns to create new images similar to the original ones.
Recently, a new type of GAN, conditional GAN (cGAN), was introduced [18]. cGAN can produce guided images with specific features, using a conditional latent variable to guide the image generation of a designated component [26]. For example, cGAN enables researchers to produce chest radiographs with anomalies from normal images. Therefore, with cGAN, it is now possible to generate images of pathologies with low incidence, thereby augmenting the dataset such that rare diseases are equally represented. The idea of conditional image generation has also been applied in image-to-image translation. By providing conditional domain information, the input image can be translated to various target domains, and the network can learn the relevant features of multiple domains through a single model.
In this study, a repertoire of realistic chest radiographs was synthesized using image-to-image translation; subsequently, the clinical competency of generated images was assessed through multi-center visual Turing test by five board-certified radiologists from three university hospitals. Furthermore, an investigation was conducted to discern whether GAN augmentation enhances convolutional neural network (CNN) performances by evaluating three CNN classifiers trained on the original and GAN-augmented datasets. Therefore, the aim of this study was to augment imbalanced medical data using generative adversarial networks and evaluate the clinical quality of the generated images via a multi-center visual Turing test.
2. Methods
This study was HIPAA compliant. Per the Office for Human Research Protections (45CFR46.102), publicly available data do not require Institutional Review Board (IRB) review. Nevertheless, Seoul National University Hospital IRB reviewed the study protocol and the need for informed consent was waived (IRB number E-2211-041-1375).
This study comprised two major tasks: first, publicly available datasets were used to train GAN to synthesize realistic chest radiographs; and second, the generated images were evaluated through 1) a multi-center visual Turing test and 2) the performance metrics of the CNN-based classifier trained on original and GAN-augmented dataset.
2.1 Image synthesis
2.1.1 Dataset.
Six publicly available chest radiograph datasets were used for both the GAN and CNN-based classifiers, namely ChestX-ray 8, MIMIC-CXR, CheXpert, JSRT, VBD, and OpenI. The details of the datasets used in this study are presented below in Table 1.
[Figure omitted. See PDF.]
Of these six datasets, one (OpenI) was reserved to construct a separate test set, while the other five were used to construct the training/validation set to train the GAN and CNN-based classifiers. To represent cardiac, pulmonary, and neoplastic pathologies, pleural effusion (pulmonary), cardiomegaly (cardiac), and lung mass (neoplastic) were used as disease labels. Images with unique labels (only one positive disease label) were obtained. The distribution of labels in the primary repertoire of raw data is presented in Table 2.
[Figure omitted. See PDF.]
The cumulative training/validation dataset consisted of 1,560 radiographs in the lung mass class, 672 in the effusion class, 260 in the cardiomegaly class, and 3,230 in the no pathological findings class. The separate test set, extracted from the OpenI dataset, consisted of 97 in the lung mass class, 96 in the effusion class, 297 in the cardiomegaly class, and 1,379 in the no pathological findings class. The distribution of labels in the training/validation/test sets is presented in Table 3.
[Figure omitted. See PDF.]
2.1.2 Data pre-processing.
Using in-house U-Net-based lung segmentation software, each image in the cumulative dataset was processed to yield a square region of interest that contained lung fields only. The cropped images were converted to 256 by 256 pixels and auto-contrasted after which they were filtered with global contrast factor (GCF) threshold before training the GAN- and CNN-based classifiers. This process is illustrated in Fig 1.
[Figure omitted. See PDF.]
2.1.3 Traditional augmentation.
To construct the first of the two augmented datasets, the original cumulative dataset was augmented using traditional augmentation methods, such as geometric rotation, flipping, cropping, and scaling. The augmentation process included random rotation between −10 and 10 degrees, random linear translation between −10 and 10 pixels, horizontal flipping, random scaling between factors 0.85 and 1.15, and random contrast between 90% and 110%. The conventional augmentation schemes employed in this study are widely used in computer vision research [17], yet have been reported to result in overfitting [19] or computational burden without performance enhancement [22]. Traditional augmentation has been reported to occasionally result in worse performance metrics [22]. The dataset was augmented such that the underrepresented positive label classes would be balanced with the ‘no findings’ class. The data distribution for the augmented sets is presented in Table 3.
2.1.4 GAN augmentation.
GAN employ two neural networks to create plausible synthetic data that retains the critical features of the original data [23]. Recent developments in GAN research have resulted in impressive levels of image generation using image-to-image translation. In image-to-image translation, the network learns to map a feature from input x onto a target output y; therefore, one can expect to generate an image with the desired disease label from a normal input image [26]. In this study, a unified conditional GAN, StarGAN v2, was used to perform GAN augmentation.
StarGAN v2 was trained to create (i) a cardiomegaly class from the normal class, (ii) an effusion class from the normal class, and (iii) a lung mass class from the normal class. As shown in Fig 2, the image styles of each class were transferred using instance normalization layers managed by the conditional encoder. The object structure and texture information were analyzed separately to generate the combined images after which the geometric information of the object was encoded using ResNet-like encoders (ResBlk) while low-level features of the texture information were extracted using style encoders. Subsequently, each AdaInResBlk block containing AdaIn modules modulated the learned geometric representation by cues received from the style encoder. To quantitatively evaluate the quality of the generated images, the differences between the two distributions in the high-dimensional feature space of an InceptionV3 classifier with a Fréchet inception distance score were measured. The augmentation process is illustrated in Fig 2, whereas the data distribution for the augmented set is presented in Table 3.
[Figure omitted. See PDF.]
2.2 Experiments
2.2.1 Visual Turing test.
Five board-certified radiologists from three different university hospitals, each with at least five years of clinical experience, independently assessed the quality of GAN-generated chest radiographs. Each radiologist was given a set of 103 chest radiographs (55 GAN-generated and 48 real) and had to decide whether each image was real or artificial using visual analysis without a time limit. Each radiologist was blinded to the composition of the test set. The mean accuracy, sensitivity, specificity, F1-score, and Matthews correlation coefficient (MCC) of the five experts were calculated. Mann-Whitney U test was used to evaluate whether the mean of the expert performance was better than random guessing. McNemar test was used to investigate whether each expert response was significantly different from random.
To assess whether the GAN-generated images were not only realistic but also clinically informative, the radiologists were further assigned to two classification tasks. For the first task, the radiologists classified 103 actual chest radiographs into four classes: effusion, cardiomegaly, lung mass, and no findings. The second task was identical to the first, except that it was conducted using GAN-generated images. The mean accuracies of the five experts were calculated for each task.
2.2.2 Training CNN classifier.
To develop and evaluate the performance of the chest radiograph classifier, four CNN networks were used, namely ResNet50, VGG16, InceptionV3, and DenseNet121. Moreover, we trained SqueezeNet and MobileNetV2 to check the performance of shallow CNN models. Each network was trained on three train/validation sets: the original dataset, the traditionally augmented dataset, and the GAN-augmented dataset. Pre-trained ImageNet models were used to customize the classifiers and for each pre-trained network, the last fully connected layer was modified to match the number of classes (four). The models were trained with a batch size of 16 and a SGD optimizer with a learning rate of 1× (10)-4. To avoid overfitting, we enabled the early stopping with patience = 10. The output layers of models includ the softmax activation function and the categorical cross-entropy loss function. The architectures of networks are described in S1 Fig in File. After training, the models were evaluated using a test set with a separate chest radiograph dataset (OpenI). The learning curves of the evaluated CNN classifiers are shown in S2 Fig in S1 File.
2.2.3 Training GAN.
StarGANv2 was trained using the Adam optimizer with momentum parameters β_1 = 0 and β_2 = 0.99 at a learning rate of 1×(10)-4 for the generator, discriminator, and style encoder, and 1× (10)-6 for the mapping network. The total loss function consists of adversarial loss, style reconstruction loss, style diversification loss, and cycle consistency loss to preserve source characteristics. The training phase required approximately 250,000 iterations (141 hours) with a batch size of 8 on a NVIDIA DGX A100×2 (122 GB RAM) GPU.
3. Results
In this section, an analysis on the quality of the GAN-generated radiographs and whether GAN augmentation enhances the CNN-based classifier performance is presented. The synthetic radiographs generated by the GAN are shown with the actual images in Fig 3.
[Figure omitted. See PDF.]
3.1 Visual Turing test
In the visual Turing test, the radiologists were asked to distinguish GAN-generated radiographs from actual images. The mean accuracy of the five radiologists was 66.2%, which was higher than that of random guessing (341/515, 66.2% vs. 257.5/515 50.0%, respectively, P = 0.006). However, in terms of identifying GAN-generated images as artificial, there was no significant difference in the sensitivity of the radiologists and random guessing (147/275, 57.8% vs. 137.5/275, 50.0%, respectively, P = .284). A McNemar test found that only three of the five radiologists (R3, R4, and R5; P = .002, 0.014, <0.001, respectively) performed better than random guessing. The average MCC value was 0.351, ranging from 0.184 to 0.549. The results presented are given in Table 4.
[Figure omitted. See PDF.]
The radiologists were then given two image sets, one with 103 actual chest radiographs and another with 103 GAN-generated radiographs, and asked to classify each image. Each dataset equally contained three disease classes (cardiomegaly, pleural effusion, lung mass), and one normal class.
Results showed that the experts performed better on GAN-generated images than actual radiographs, (average accuracy 73.7% vs 58.6%, P<0.01), with the mean accuracy differing significantly in larger pathology classes (cardiomegaly and pleural effusion) but not in smaller lesions (lung mass) or normal images. The resulting metrics for the multiclass classification task can be found in Table 5. A possible explanation for this interesting gap in accuracy is presented in the discussion section.
[Figure omitted. See PDF.]
3.2 CNN performance
Four CNN-based classifiers, each with a ResNet50, VGG16, InceptionV3, and DenseNet121 backbone, were trained and evaluated. Each classifier was trained on three different train/validation sets, namely the original cumulative train/validation set before augmentation, traditional augmentation (TA), and GAN augmentation. The augmentation amplitudes are presented in Table 3.
The classifiers were then tested on a test set comprising a separate dataset (OpenI) to determine the external validity of the models. The ResNet50 model outperformed the other two classifiers, achieving 83.9% accuracy after GAN augmentation. This model also benefited the most from GAN augmentation, with a 6.6% increase in performance from 78.7% to 83.9%. After GAN augmentation, the accuracy of this model was comparable to that of radiologists. The performance metrics for the multiclass classification are listed in Table 6 and the performance of shallow CNN models could be found in S1 and S2 Tables in S1 File.
[Figure omitted. See PDF.]
In a binary classification task, wherein the classifier determined whether an image was normal (without any pathologic findings) or abnormal (with any pathologic finding), the effect of GAN augmentation was more evident. The ResNet50 model also outperformed the other two classifiers, with an AUROC of 0.871. This model benefited the most from GAN augmentation, increasing its AUROC by 8.6% from 0.802 to 0.871. The performance metrics of binary classification are listed in Table 7 and the receiver operating characteristic (ROC) curves of the ResNet50 model are shown in Fig 4.
[Figure omitted. See PDF.]
(A) ResNet50, (B) VGG16, (C) InceptionV3, and (D) DenseNet121.
[Figure omitted. See PDF.]
4. Discussion
This study addressed two distinct questions: firstly, whether the generated radiographs deemed realistic to the medical experts, and secondly, whether the GAN-based data augmentation improved the performance of the in-house medical AI.
An open-source, heterogeneous group of chest radiograph datasets was used to ensure that the proposed GAN model was robust to different acquisition parameters of various X-ray machines. As shown in Fig 3, the model produced a realistic repertoire of chest radiographs. Considering that data heterogeneity may result in the production of imperfect chest radiographs, the quality of the generated images was evaluated by asking radiologists working at university hospitals with at least five years of clinical experience to determine whether a given image was real or AI-generated.
As shown in Table 4, radiologists could distinguish GAN-generated images from real images with a 66.2% accuracy. Although this accuracy is better than random guessing, the sensitivity, or the accuracy of correctly telling a GAN-generated image apart, was not significantly different from random guessing. Furthermore, the responses from two of the five radiologists were not significantly different from random guessing, according to the McNemar test. This illustrates that the synthesized images appeared realistic to medical experts.
Subsequently, an evaluation was conducted on the performance of medical experts on multiclass classification of pathologies on both real and GAN-generated images. Here, the average accuracy was 58.6% for actual images and 73.7% for GAN-generated images (Table 5). The apparent gap in performance is perplexing and raises the question of how the radiologists performed better on the AI-generated image set.
As shown in Table 5, the radiologists diagnosed pleural effusion and cardiomegaly significantly better. However, diagnostic performance on lung mass and normal images did not differ significantly across the two groups. This discrepancy may be explained by the lesion size. Lung masses are small masses within the lung parenchyma that are less than 3 cm in size, and pleural effusion is the collection of fluid at the bottom of the thorax, seen as blunted costophrenic angles on radiographs. Cardiomegaly can be detected by the enlargement of heart shadow over 60% of the thoracic width. Cardiomegaly and pleural effusion are both large lesions and present distinct features that GAN can detect. Because the GAN generator learns to replicate distinct features of an image to convince the discriminator that the image is real, large features that strongly represent the real pathology may be reinforced, making the GAN generate radiographs with pathologies easier to classify.
Tables 6 and 7 show that GAN augmentation enhanced CNN classifier performance significantly. This result is noteworthy because it showed that GAN augmentation of existing clinical datasets could improve the medical AI aided diagnostic performance. In all of the CNN classifier models, training on GAN augmented dataset improved medical AI performance. Because CNN based model performances rely on the amount and quality of trainable data, the present result proposes a possibility of improving existing computer aided diagnosis programs. If medical doctors were to use medical AI aided diagnosis algorithms, they would benefit from AI’s that show better performance by training on GAN augmented dataset. Additionally, the shallow CNN models also show increased performance with GAN augmented dataset (S1 and S2 Tables in S1 File), despite the relatively small parameters. These results show that GAN augmentation could have a potentially meaningful role in future lightweight CNN model research.
These findings suggest two distinct strengths of GAN augmentation for chest radiographs. Firstly, GAN-generated images retain most clinical features and are difficult to distinguish from real images, even by clinical experts. Secondly, while GAN-generated images are similar to real images, they also emphasize important clinical attributes, such as pathological radiological signs. While these strengths may serve to enhance the overall performance of medical AI, they can also serve to reproduce medical data with distinct features for human educational purposes. In their 2018 research, Finlayson et al. proposed a similar GAN-based training tool for medical education [31]. GAN images may serve as effective augmentation, especially in low-data regimes such as medical fields.
This study has limitations. Firstly, the chest radiographs used were 256 × 256 pixels. In clinical settings, radiologists have access to powerful diagnostic software and images that are at least 2048 × 2048 pixels in size, however, this is only a technical limitation. The resolution of generated radiographs can be expanded with more powerful computing resources. Meanwhile, the medical experts achieved competent classification accuracy even with the current resolution. Another limitation is that the research employed only public-domain datasets. The public datasets employed in this study are prone to data ambiguity because many of them used classification algorithms instead of manual annotation during labeling. Samples from the datasets, however, have been cross-checked by human experts to ensure that they are of competent clinical quality.
This research used GAN to produce chest radiographs that not only improved AI performance but were also convincing as real images to board-certified radiologists. From here, there are plans to extend the research to other imaging modalities with imbalanced datasets, such as rare brain tumors or pediatric cancer.
5. Conclusion
The proposed GAN produced realistic chest radiographs that appeared realistic to experienced radiologists; in the visual Turing test, there was no significant difference in the sensitivity of identifying GAN images as artificial between radiologists and random guessing. Radiologists effectively classified chest pathologies with the synthesized radiographs, suggesting that the GAN images contained adequate clinical information for diagnosis. The multi-center visual Turing test found that GAN tends to emphasize larger lesions better than smaller pathologies. To the best of our knowledge, this finding has not been reported before in medical image analysis using GAN. CNN based classifiers rely on the amount and quality of training data, and therefore the result of the present study is noteworthy because it showed that GAN augmentation could effectively augment training data for medical AI. Augmentation through the synthesized images significantly enhanced CNN classifier performance, providing a successful means to overcome data imbalance in medical image analysis.
Supporting information
S1 File.
https://doi.org/10.1371/journal.pone.0279349.s001
(DOCX)
Citation: Myong Y, Yoon D, Kim BS, Kim YG, Sim Y, Lee S, et al. (2023) Evaluating diagnostic content of AI-generated chest radiography: A multi-center visual Turing test. PLoS ONE 18(4): e0279349. https://doi.org/10.1371/journal.pone.0279349
About the Authors:
Youho Myong
Contributed equally to this work with: Youho Myong, Dan Yoon
Roles: Conceptualization, Data curation, Writing – original draft, Writing – review & editing
Affiliations: Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, Republic of Korea, Department of Rehabilitation Medicine, Seoul National University Hospital, Seoul, Republic of Korea
Dan Yoon
Contributed equally to this work with: Youho Myong, Dan Yoon
Roles: Conceptualization, Data curation, Software, Visualization, Writing – original draft, Writing – review & editing
Affiliation: Interdisciplinary Program in Bioengineering, Seoul National University Graduate School, Seoul, Republic of Korea
Byeong Soo Kim
Roles: Conceptualization, Data curation, Formal analysis, Software, Writing – review & editing
Affiliation: Interdisciplinary Program in Bioengineering, Seoul National University Graduate School, Seoul, Republic of Korea
Young Gyun Kim
Roles: Conceptualization, Data curation, Formal analysis, Software, Visualization, Writing – review & editing
Affiliation: Interdisciplinary Program in Bioengineering, Seoul National University Graduate School, Seoul, Republic of Korea
ORICD: https://orcid.org/0000-0003-1231-9097
Yongsik Sim
Roles: Validation, Writing – review & editing
Affiliation: Department of Radiology, Severance Hospital, Yonsei University College of Medicine, Seoul, Republic of Korea
ORICD: https://orcid.org/0000-0003-2711-2793
Suji Lee
Roles: Validation, Writing – review & editing
Affiliation: Department of Radiology, Severance Hospital, Yonsei University College of Medicine, Seoul, Republic of Korea
Jiyoung Yoon
Roles: Validation, Writing – review & editing
Affiliation: Department of Radiology and Research Institute of Radiological Science, Yonsei University College of Medicine, Seoul, Republic of Korea
Minwoo Cho
Roles: Conceptualization, Writing – review & editing
E-mail: [email protected] (SK); [email protected] (MC)
¶‡ SK and MC also contributed equally as co-corresponding authors.
Affiliations: Transdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Seoul, Republic of Korea, Medical Big Data Research Center, Seoul National University College of Medicine, Seoul, Republic of Korea, Biomedical Research Institute, Seoul National University Hospital, Seoul, Republic of Korea
Sungwan Kim
Roles: Conceptualization, Funding acquisition, Resources, Supervision, Writing – review & editing
E-mail: [email protected] (SK); [email protected] (MC)
¶‡ SK and MC also contributed equally as co-corresponding authors.
Affiliations: Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, Republic of Korea, Interdisciplinary Program in Bioengineering, Seoul National University Graduate School, Seoul, Republic of Korea, Institute of Bioengineering, Seoul National University, Seoul, Republic of Korea
ORICD: https://orcid.org/0000-0002-9318-849X
1. Wu JT, Wong KCL, Gur Y, Ansari N, Karargyris A, Sharma A, et al. Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents. JAMA Netw Open. 2020;3(10):e2022779. Epub 20201001. pmid:33034642
2. Jacobi A, Chung M, Bernheim A, Eber C. Portable chest X-ray in coronavirus disease-19 (COVID-19): A pictorial review. Clinical imaging. 2020;64:35–42. pmid:32302927
3. Serrano CO, Alonso E, Andrés M, Buitrago N, Vigara AP, Pajares MP, et al. Pediatric chest x-ray in covid-19 infection. European Journal of Radiology. 2020;131:109236. pmid:32932176
4. Cozzi D, Albanesi M, Cavigli E, Moroni C, Bindi A, Luvarà S, et al. Chest X-ray in new Coronavirus Disease 2019 (COVID-19) infection: findings and correlation with clinical outcome. La radiologia medica. 2020;125:730–7. pmid:32519256
5. Minaee S, Kafieh R, Sonka M, Yazdani S, Soufi GJ. Deep-covid: Predicting covid-19 from chest x-ray images using deep transfer learning. Medical image analysis. 2020;65:101794. pmid:32781377
6. Gavelli G, Giampalma E. Sensitivity and specificity of chest x‐ray screening for lung cancer. Cancer. 2000;89(S11):2453–6.
7. Bradley S, Abraham S, Grice A, Lopez RR, Wright J, Farragher T, et al. Sensitivity of chest X-ray for lung cancer: systematic review. British Journal of General Practice. 2018;68(suppl 1).
8. Bradley SH, Abraham S, Callister ME, Grice A, Hamilton WT, Lopez RR, et al. Sensitivity of chest X-ray for detecting lung cancer in people presenting with symptoms: a systematic review. British Journal of General Practice. 2019;69(689):e827–e35. pmid:31636130
9. Stirenko S, Kochura Y, Alienin O, Rokovyi O, Gordienko Y, Gang P, et al., editors. Chest X-ray analysis of tuberculosis by deep learning with segmentation and augmentation. 2018 IEEE 38th International Conference on Electronics and Nanotechnology (ELNANO); 2018: IEEE.
10. Smith DL, Grenier JP, Batte C, Spieler B. A Characteristic Chest Radiographic Pattern in the Setting of the COVID-19 Pandemic. Radiol Cardiothorac Imaging. 2020;2(5):e200280. Epub 20200903. pmid:33778626
11. Cho Y, Kim Y-G, Lee SM, Seo JB, Kim N. Reproducibility of abnormality detection on chest radiographs using convolutional neural network in paired radiographs obtained within a short-term interval. Scientific Reports. 2020;10(1):1–11.
12. Miller WD, Peek ME, Parker WF. Scarce Resource Allocation Scores Threaten to Exacerbate Racial Disparities in Health Care. Chest. 2020;158(4):1332–4. Epub 20200522. pmid:32450243
13. Junhewk K. Setting priorities for medical resource allocation of emergency treatment and vaccination for COVID-19 in South Korea. Bio, Ethics and Policy. 2020.
14. Kirkpatrick JN, Hull SC, Fedson S, Mullen B, Goodlin SJ. Allocating Scarce Resources and Triaging Patients during the COVID-19 Pandemic. Journal of the American College of Cardiology. 2020.
15. Rajpurkar P, Irvin J, Ball RL, Zhu K, Yang B, Mehta H, et al. Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS medicine. 2018;15(11):e1002686. pmid:30457988
16. Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Medical image analysis. 2017;42:60–88. pmid:28778026
17. Bria A, Marrocco C, Tortorella F. Addressing class imbalance in deep learning for small lesion detection on medical images. Comput Biol Med. 2020;120:103735. Epub 20200401. pmid:32250861.
18. Yi X, Walia E, Babyn P. Generative adversarial network in medical imaging: A review. Medical image analysis. 2019;58:101552. pmid:31521965
19. Balachandar N, Chang K, Kalpathy-Cramer J, Rubin DL. Accounting for data variability in multi-institutional distributed deep learning for medical imaging. J Am Med Inform Assoc. 2020;27(5):700–8. pmid:32196092
20. Wang X, Peng Y, Lu L, Lu Z, Bagheri M, Summers RM, editors. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. Proceedings of the IEEE conference on computer vision and pattern recognition; 2017.
21. Johnson AE, Pollard TJ, Berkowitz SJ, Greenbaum NR, Lungren MP, Deng C-y, et al. MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports. Scientific data. 2019;6(1):1–8.
22. Shorten C, Khoshgoftaar TM. A survey on image data augmentation for deep learning. Journal of Big Data. 2019;6(1):1–48.
23. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. Advances in neural information processing systems. 2014;27.
24. Creswell A, White T, Dumoulin V, Arulkumaran K, Sengupta B, Bharath A. Generative adversarial networks: an overview. IEEE Signal Process Mag 35 (1): 53–65. 2018.
25. Wang K, Gou C, Duan Y, Lin Y, Zheng X, Wang F-Y. Generative adversarial networks: introduction and outlook. IEEE/CAA Journal of Automatica Sinica. 2017;4(4):588–98.
26. Choi Y, Uh Y, Yoo J, Ha J-W, editors. Stargan v2: Diverse image synthesis for multiple domains. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2020.
27. Irvin J, Rajpurkar P, Ko M, Yu Y, Ciurea-Ilcus S, Chute C, et al., editors. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. Proceedings of the AAAI conference on artificial intelligence; 2019.
28. Shiraishi J, Katsuragawa S, Ikezoe J, Matsumoto T, Kobayashi T, Komatsu K-i, et al. Development of a digital image database for chest radiographs with and without a lung nodule: receiver operating characteristic analysis of radiologists’ detection of pulmonary nodules. American Journal of Roentgenology. 2000;174(1):71–4. pmid:10628457
29. Nguyen HQ, Lam K, Le LT, Pham HH, Tran DQ, Nguyen DB, et al. VinDr-CXR: An open dataset of chest X-rays with radiologist’s annotations. arXiv preprint arXiv:201215029. 2020.
30. Demner-Fushman D, Kohli MD, Rosenman MB, Shooshan SE, Rodriguez L, Antani S, et al. Preparing a collection of radiology examinations for distribution and retrieval. Journal of the American Medical Informatics Association. 2016;23(2):304–10. pmid:26133894
31. Finlayson SG, Lee H, Kohane IS, Oakden-Rayner L. Towards generative adversarial networks as a new paradigm for radiology education. arXiv preprint arXiv:181201547. 2018.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 Myong et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Background
Accurate interpretation of chest radiographs requires years of medical training, and many countries face a shortage of medical professionals to meet such requirements. Recent advancements in artificial intelligence (AI) have aided diagnoses; however, their performance is often limited due to data imbalance. The aim of this study was to augment imbalanced medical data using generative adversarial networks (GANs) and evaluate the clinical quality of the generated images via a multi-center visual Turing test.
Methods
Using six chest radiograph datasets, (MIMIC, CheXPert, CXR8, JSRT, VBD, and OpenI), starGAN v2 generated chest radiographs with specific pathologies. Five board-certified radiologists from three university hospitals, each with at least five years of clinical experience, evaluated the image quality through a visual Turing test. Further evaluations were performed to investigate whether GAN augmentation enhanced the convolutional neural network (CNN) classifier performances.
Results
In terms of identifying GAN images as artificial, there was no significant difference in the sensitivity between radiologists and random guessing (result of radiologists: 147/275 (53.5%) vs result of random guessing: 137.5/275, (50%); p = .284). GAN augmentation enhanced CNN classifier performance by 11.7%.
Conclusion
Radiologists effectively classified chest pathologies with synthesized radiographs, suggesting that the images contained adequate clinical information. Furthermore, GAN augmentation enhanced CNN performance, providing a bypass to overcome data imbalance in medical AI training. CNN based methods rely on the amount and quality of training data; the present study showed that GAN augmentation could effectively augment training data for medical AI.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer