Introduction
The latest paradigm shift of machine learning towards deep learning (DL) is spurred by its success on many longstanding computer vision tasks. This has motivated exploration of DL in wide-ranging medical applications from disease detection [1] to segmentation [2]. Since DL is a data-driven framework, its success is contingent on the abundance of training data with expert annotations. Acquisition of expert annotations has always been difficult in the medical domain given the tedium of the task and the priority patient care takes over the annotation task. Data augmentation (via geometric transformations) for robust training is a popular solution adopted by the computer vision community. However, this has limitations in the medical domain as it does not introduce any real variability that is essential for robust learning of abnormalities, normal anatomy and so on. We examine this problem of sparsity of annotated data and explore two different avenues for solutions: (i) crowdsourcing and (ii) synthesis.
Crowdsourcing has been shown to be reliable [3–5] and useful to train classifiers [6]. Annotations have been crowdsourced from fundus images, endoscopy and magnetic resonance imaging (MRI) of brain [3–5] for image level classification and reference correspondence. They have also been used to segment a surgical instrument from laparoscopic images [6]. An active learning framework with crowdsourcing serves to reduce the burden on the crowd as it allows only low confident samples predicted by a model to be given to the crowd. Atlas forests were updated in [7] based on crowd refined annotations (on instrument boundary) to generate a new atlas. A convolutional neural network (CNN) was trained in [8] and the crowdsourced mitosis candidates (in a patch of size ) were merged with an aggregation layer for updating the model. Crowd annotations are inherently noisy and hence merging them to derive a single ground truth (GT) is a key issue. Methods ranging from simple majority voting (MV) [6] to a stochastic modelling using expectation maximisation [4] and introducing an aggregation layer in a CNN [8] have been proposed in literature.
A second avenue that is free of human annotation is the synthesis route. Image synthesis has been attempted to generate digital brain phantoms [9] and whole retinal images [10] using complex modelling. These have been aimed at aiding denoising, reconstruction and segmentation solutions. Recently, simulation of brain tumours in MR images [11] has also been explored to aid computer aided diagnosis (CAD) algorithm development.
With the advent of DL, modelling of complex structures and synthesising images has become easier with a class of neural networks called generative adversarial networks or GAN [12]. GAN is an architecture composed of two networks, namely, a generator and a discriminator. Functionally, the generator synthesises images from noise while the discriminator differentiates between real and synthetic images. GAN have recently been explored for a variety of applications: detection of brain lesions [13], predicting computed tomography from MRI images [14], synthesising normal retinal images from vessel mask [15], segmenting anatomical structures such as vessels [16] and optic disc/cup [17].
In this paper, we take up the problem of DR lesion detection from colour fundus images and explore the use of the aforementioned two avenues to aid the development of robust CAD solution. Our contribution is three-fold:
We consider crowdsourcing as an independent (of model learning) activity and propose a scheme wherein only regions of interest (ROIs) are marked by the crowd to reduce the burden. A solution for merging crowd annotations is proposed based on assigning a reliability factor (RF) for each subject of the crowd. This leverages the abundant availability of image-level annotations to assess the subjects.
We propose a GAN for generating images with pathologies in a controlled manner.
Finally, we illustrate how a heterogeneous mixture of annotations derived from experts, crowd and through synthesis can address the data sparsity problem.
Methods
As a part of the pre-processing step, given retinal images are corrected for non-uniform illumination using luminosity and contrast normalisation [18].
Crowdsourcing annotations
Subjects and tasks
A total of 11 engineering students volunteered to be ‘crowd’ members/subjects. Four of these were familiar with fundus images () while the rest were not familiar with any medical images (). The task given to the crowd subjects was twofold: (i) determine whether the given image is normal/abnormal and (ii) if abnormal, mark the ROI containing HE. A freehand annotation tool (Paint.Net [. ]) was provided for the task. Fig. 1 shows a screenshot of the annotation tool.
[IMAGE OMITTED. SEE PDF]
Materials
Hundred images were given to each subject. Of the 100, 6 images were from DIARETDB1 [19] which provides ROI markings from 4 experts. The remaining 94 (70 with HE and 24 with no lesions) were from MESSIDOR [20] which provides annotations only at the image level. Thus, each of the 100 images that was selected for crowdsourcing has a label (normal/abnormal) and only 6 have additional information about locations of abnormalities. A sample image with HE is shown in Fig. 2 along with relevant landmarks.
[IMAGE OMITTED. SEE PDF]
Processing crowd annotations
Our crowdsourcing exercise produced 11 annotations for each of the 100 images. These need to be integrated to derive a single annotation for every image. Merging the annotations via simple MV is likely to produce noisy annotation. Hence, a more elaborate procedure was designed for merging and GT derivation. This began with assigning a RF to every subject i, shown in Fig. 3. Ideally, RF should rely on two factors: experience and performance of a subject. Experience can be determined via explicit queries. Performance needs to be assessed preferably by benchmarking against experts. A scheme was designed to reward a subject for good performance at both local ROI level (based on performance on the six images from DIARETDB1) and image level (based on known labels of all 100 images). A score is given for each of these factors and the final RF is computed as a weighted sum of these scores. The RF for the i th subject was defined as
[IMAGE OMITTED. SEE PDF]
Scoring performance at an image level: The crowd annotation for an image is binary (normal or abnormal) which is unlike the expert annotation for MESSIDOR. The latter encodes the location of HE (standard grading [20]): 0 indicating a normal image, 1 if the lesions are outside a circular region (of diameter equal to optic disc) surrounding the macula and 2 if they are inside this circular region. Hence, we assign a score to a subject not only based on correct labelling of normal images but rewarding them when their ROI is in the correct zone. is designed to be based on the true positive rate (TPR) and false positive rate (FPR) () for each subject. The ROI location of an i th subject is compared with the zonal labels (j) from MESSIDOR and the score is calculated as follows:
DNN for aggregation of crowd annotations
We propose an alternate strategy to aggregate crowd annotations using DNN to train different models with different crowd annotations as GT. The performance of the subject is assessed based on the model performance on images which have local-level and image-level markings from the expert. In this strategy, the weights () are assigned by the DNN to different factors based on the performance of the subjects.
In this approach, we chose U-net to train the models. Let be a subject and be the local annotation (image level annotation) given by the subject i on image j. Here, as there are 11 subjects and as 70 abnormal images from MESSIDOR and 6 images from DIARETDB1 are considered for training. Each U-net () is trained to detect HEs using the above 76 images for training and the corresponding crowd annotations as GT. As there are 11 subjects we obtain a total of 11 U-net models. Now, each of the U-net model is tested on DMED and DRiDB images to obtain pixel-wise classification. The sensitivity (SN) and positive predictive value (PPV) values are calculated for each model by comparing against the local GT marked by experts. Further, the pixel-level annotations are converted to image-level annotations to asses performance at image level. The RF for each subject is given based on these values as
GAN for synthesis of retinal images with pathologies
A second route we explore for deriving annotations is synthesis using a GAN made of a discriminator and a generator. A GAN learns a model as follows: the discriminator iteratively reduces its misclassification error by more accurately classifying the real and synthetic images while the generator aims to deceive the discriminator by producing more realistic images. GAN-based synthesis of normal retinal images has been demonstrated in [15] (from a vessel mask) with a single U-net for the generator and a five-layer CNN for the discriminator. The U-net architecture consists of a contracting and an expansive path. The contracting path is similar to a typical CNN architecture, whereas in the expanding path, max-pooling is replaced by up-sampling. There are skip connections between contracting and expanding paths to ensure localisation. The U-net is modified in terms of the number of filters at each convolutional layer. The number of filters at each stage are reduced to half to simplify computations.
Our interest is in the synthesis of images with HE to serve as exemplars for different stages of DR, which is based on the locations and density of HE. Hence, we designed a GAN architecture (shown in Fig. 4) with a generator consisting of two parallel networks: one with a vessel mask as input and another with a lesion mask as input. The output of the networks, based on the U-net architectures, are merged and fed to a third U-net architecture which generates the whole retinal image with lesions. The generator thus maps from vessel () and lesion () masks to a retinal image () using a mapping function. A five-layer CNN as in [15] is used for the discriminator to distinguish between the real and synthetic sets of images, with each set consisting of vessel and lesion masks along with retinal images.
[IMAGE OMITTED. SEE PDF]
The overall loss function that is to be optimised is chosen as a weighted combination of three loss functions: , and as defined below in (7)–(10) to produce sharp and realistic images.
(i) The adversarial loss function is defined as
(ii) The structure similarity (SSIM) [21] index is useful in quantitatively measuring the structural similarity between two images (r, G (v, l)). It also has been shown to perform well for reconstruction and generation of visually pleasing images
(iii) The loss function is used mainly to reduce artefacts and blurring and is defined as
CAD solution for hard exudate detection
The U-Net [22] was chosen to build a CAD solution for HE detection. This solution will be referred to as CADH. A standard architecture was chosen as the aim is to demonstrate that the crowdsourced annotations and synthetic images (generated by our proposed GAN) are reliable resources in training even a basic U-net (Fig. 5).
[IMAGE OMITTED. SEE PDF]
Preprocessing
Fundus images suffer from non-uniform illumination due to image acquisitions, camera limitations and so on. This is corrected using luminosity and contrast normalisation [18]. The optic disc region in every image is masked out and inpainted. Fundus extension is applied to remove the black mask region and all images are normalised to have zero mean and unit variance.
Data augmentation
Data augmentation is done by applying random transformations to the images. This included random rotation between and , random translation in vertical/horizontal directions in the range of 50 pixels and random horizontal/vertical flips. For fairness, the number of images used for data augmentation are chosen to be equal to crowdsourced images/synthetic images.
Implementation details
Datasets
Four public datasets, namely, DRiDB [23], DMED [24], MESSIDOR and DIARETDB1 were used in this work. DMED has pixel-level annotations (from one expert) whereas DIARETDB1 (four experts) and DRiDB (one expert) have ROI markings. The consensus marking of three experts was used to derive a binary mask in the case of DIARETDB1. The obtained binary mask was overlapped on the image and thresholded to get pixel-level lesion mask. Images from all datasets were cropped and resized 512 × 512 before feeding them to GAN or CADH.
DNN for aggregation of crowd annotations
The training of each U-net consisted of 70 abnormal images from MESSIDOR and 6 abnormal images from DIARETDB1 given to the crowd for annotation. After augmentation it accounts to a total of 152 images for training with the corresponding crowd annotations as GT. The testing consists of 84 abnormal images, 31 from DRiDB and 53 from DMED.
DNN for HE detection
Testing of CADH was done with DIARETDB1 (42 images). Training of CADH was done with the following datasets: expert annotations of DRIDB (31 images) and DMED (53 images); crowdsourced annotations of MESSIDOR (70 images). Additionally, annotations from synthetic images generated from GAN were also used which is described next.
Training data for GAN
Training of the GAN requires both lesion and vessel masks. The lesion mask for the training data is available from experts but vessels mask are available only for DRiDB. It is tedious and time consuming to mark the vessels in each of the retinal images. Hence, vessel masks were derived using a vessel segmentation method [25] which has proved to be robust to pathologies.
The synthetic retinal images were generated using GAN as follows. The required vessel and lesion masks were sourced from images selected randomly from DMED and DRiDB. The lesion masks were modified using the same random transformations such as flipping the lesions sector wise, flipping horizontally and vertically, rotations and translations. Retinal images containing HE are graded with severity levels as in [20]. The lesions masks were derived to provide exemplars for each level using these rules. The position of lesions in each category was maintained by masking out few lesions or adding new lesions from another lesion mask randomly.
Computing details
The models were implemented in Python using Keras with Theano as backend and trained on a NVIDIA GTX 970 GPU, 4 GB RAM. Training was done with random initialised weights for 2000 epochs by minimising the loss function using Adam optimiser. For model parameters, learning rate was initialised to for GAN and for CADH. A batch size of 4 was considered for both the cases and other parameters were left at default values. Class weights were outlined as the inverse ratio of the number of positive samples to negative samples and modified empirically.
Evaluation metrics
Several experiments were conducted to assess the relative merits of crowdsourcing and synthesis of annotated data for training CADH. The merit was determined based on the HE detection performance.
Crowd annotations: Crowdsourced annotations were assessed with TPR, FPR and accuracy as evaluation metrics. As the image-level labels available from the experts is for three classes (labelled i: 0, 1 and 2), TPR, FPR and accuracy were calculated as follows:
Aggregation of crowdsourced annotations: The label aggregation is assessed using TPR, FPR in respective zones and also accuracy. The statistical significance of the aggregated labels against baseline is also calculated by using p -value.
Assessment of generated synthetic images: The synthesised images were evaluated quantitatively and qualitatively (five sample results are shown in Fig. 6). The mean and standard deviation of the score described in [26] was computed over all images (42 abnormal) in DIARETDB1.
[IMAGE OMITTED. SEE PDF]
Assessment of CADH: The performance of CADH was evaluated using SN and PPV which are defined as follows: and . To evaluate against the given local annotations by experts, the pixel-wise classification was converted to region-wise detection by applying connected component analysis and requiring at least 50% (but not exceeding more than 150%) overlap with manually marked regions to identify true positive detections (TP); else it is false positive (FP). If a region is marked by the expert but was not detected by the model then it is a false negative (FN). The area under the SN versus PPV curve (AUC) is also taken as a measure of performance.
Results
Crowdsourced annotations
Crowdsourced data: The average time taken by subjects to mark ROI for 100 images was around 90 min. The task was conducted in two sessions of 50 images each. Hence, a total of 1100 markings were obtained in a span of two days. Sample retinal image and respective 11 crowd annotations are shown in Fig. 7. The annotation performance is presented as a box plot in Fig. 8 for the three zonal labels/classes. The mean accuracy obtained was 70% and the class-wise TPR/FPR figures were: 89.6%/6.9% for normal/class 0; 80.7%/11.29% for class 1 and 77.69%/10.7% for class 2. These indicate that the crowd is good at correctly identifying normal images and detects HE in the large zone 1 more accurately than much smaller ±zone 2 (size of optic disc) suggesting a bias towards the larger zone. Since lesions in zone 2 require immediate referral, urging subjects to scrutinise this zone may be advisable.
[IMAGE OMITTED. SEE PDF]
[IMAGE OMITTED. SEE PDF]
Aggregation of labels: The impact of the factors that contribute to the proposed RF (see (1)) was studied by setting . The obtained TPR and FPR are listed in Table 1. The baseline is taken as MV in the following discussion. When only image-level performance is considered for the RF, there is a 10% improvement in accuracy which is boosted to 22% with the addition of performance at the local level. This is noteworthy as performance at local level is known only for 6% of the images given to the crowd. The aggregation result of different annotations considering different factors is shown in Fig. 9. The aggregation of annotations using DNN gave an overall accuracy of 97% with and .
Table 1 Assessment of the scheme for label aggregation
Accuracy | |||||||
I | 100 | 1.7 | 87.9 | 3.3 | 90.9 | 6.6 | 86.2 |
I + L | 100 | 15.3 | 100 | 16.6 | 93.9 | 0 | 97.8 |
I + E | 100 | 7.57 | 97 | 0 | 87.9 | 0 | 90 |
I + L + E | 100 | 6 | 100 | 0 | 87.9 | 0 | 91.8 |
MV | 89.3 | 3.5 | 78.8 | 5.2 | 91 | 13.5 | 75.7 |
[IMAGE OMITTED. SEE PDF]
The statistical significance of different aggregations as compared to the baseline MV is shown in Table 2. This shows that I + L is statistically significant compared to other combinations. The p -value is also reported for different pair of combinations to estimate the importance of each factor (Table 3). We can infer that I + L can be considered as the alternate hypothesis, rejecting I + E and I + L + E. Experience does not seem to be beneficial for this experiment as performance suffers and also the statistical significance is less. This may be due to the fact that crowd was made of students and hence experience is really not meaningful.
Table 2 Statistical significance value for label aggregation
Baseline | Combination | p -value |
MV | I | 0.3404 |
MV | I + L | 0.0527 |
MV | I + L + E | 0.4127 |
MV | I + E | 0.4345 |
Table 3 Statistical significance value for different combination pairs of label aggregation
Baseline | Combination | p -value |
I | I + L | 0.067 |
I | I + L + E | 0.7072 |
I | I + E | 0.70 |
I + E | I + L | 0.0358 |
I + L + E | I + E | 0.442 |
I + L + E | I + L | 0.0595 |
Synthetic image generation (GAN)
Two sample retinal images (with HE) generated by the proposed GAN model are shown in Fig. 6. The first two columns show the vessel and lesion masks given as input to the GAN. The next two columns show the synthesised and the corresponding real images. The synthesised images appear realistic yet differ from the real images in terms of background colour, texture and illumination. Lesion locations are roughly similar but sizes are different as lesion masks are not results of exact segmentations of lesions.
The mean/standard deviation of computed over all images (42) with pathologies in DIARETDB1 are 0.074/0.017 and over all the synthetic images generated from vessel and lesion mask from DIARETDB1 is 0.082/0.02. These values are nearly equal indicating synthetic and real images are similar.
CADH for hard exudate detection
The utility of crowdsourcing and synthesising annotations for CAD development was tested separately with four CADH models derived by training with different training sets. Denoting the set of real images with expert annotations as E, the crowdsourced annotations as C (Experiment-1) and the set of synthetic images generated by GAN with the corresponding lesion masks as S (Experiment-2), the variants of the training set considered were: (i) only E, (ii) E with data augmentation (E + A), (iii) E and C (S), (iv) E, C (S) with data augmentation (E + (C (S)) + A). For an SN of 90%, the computed PPV and AUC values are reported in Table 4 for Experiment-1 and Experiment-2. The SN versus PPV curve is shown in Fig. 10 a for Experiment-1 and Fig. 10 b for Experiment-2. The DNN output for HE detection and the corresponding expert annotations is shown in Fig. 11.
Table 4 HE detection performance with crowdsourced data and synthetic data for training (E, expert; A, augmentation; S, synthetic; and C, crowd)
Source of training data (no. of images) (T -total no. of images) | SN, % | PPV, % | AUC |
E (84) (T -84) | 90 | 60.3 | 0.750 |
E (84) + A (70) (T -154) | 89.8 | 61.6 | 0.765 |
E (84) + S (70) (T -154) | 90 | 64 | 0.821 |
E (84) + C (I + L) (70) (T -154) | 90.1 | 71.5 | 0.869 |
E (84) + C (MV) (70) + A (154) (T -308) | 90 | 84.6 | 0.839 |
E (84) + C (I) (70) + A (154) (T -308) | 90 | 85 | 0.879 |
E (84) + S (70) + A (154) (T -308) | 90 | 82 | 0.894 |
E (84) + C (I + L) (70) + A (154) (T -308) | 90.1 | 90.4 | 0.932 |
E (84) + C (70) + S (70) + A (224) (T -448) | 89.8 | 92.0 | 0.942 |
E (84) + C (70) + S (140) + A (294) (T -588) | 90 | 92.8 | 0.95 |
E (84) + C (140) + S (140) + A (364) (T -728) | 89.7 | 93.4 | 0.956 |
[IMAGE OMITTED. SEE PDF]
[IMAGE OMITTED. SEE PDF]
Based on the figures in the table, we observe the following. Data augmentation helps improve the AUC and PPV by about 2% each, whereas, inclusion of C (S) helps improve AUC by 15.8 (9.5)% and PPV by 18.5 (6.7)%. Finally, when the expert and crowdsourced annotations are augmented and added to the training set, there is a further improvement in AUC by 24.5 (19)% and in PPV by 50 (37)%.
Setting PPV to 70% results in SN values ranging from 70 to 96% (Fig. 10); which is a 37% improvement in SN (similar level of improvement as that of PPV when SN is set constant at 90%). Thus, we conclude that C (S) is very effective in boosting the performance of CADH.
Most recent approaches for HE detection report at the image level (normal or has HE) rather than at a local level. The exception is [27] where a DL-based approach is reported to have an score of 0.78 with SN and PPV of 78% each on 50 images from DRIDB dataset.
Conclusion
In this paper, we have explored two options to address sparsity of annotated medical data which is critical for developing DL-based CAD solutions. Crowdsourcing is an alternate source of annotation, but can be effective only if measures are taken to improve the reliability of annotations. The proposed RF concept aid weighted merging of annotations with good performance rewarded with a higher weight; it was shown that it is possible to assess the ‘goodness’ of a performance with very little cost (getting the crowd to annotate a small set of images previously annotated by experts). GAN-based synthesis is another alternative. A GAN solution was proposed to generate the retinal images with HE using vessel and lesion masks. This approach gives user a greater control as retinal images can be synthesised with any type of severity, by providing the corresponding lesion mask.
Our experiment results indicate that overall, the crowdsourced annotations and synthetic data (by themselves or in combination) are reliable for developing DL-based CAD solutions. Specifically, annotations via crowdsourcing data proved to be more effective than via synthesis (a PPV of 90% versus 82% for SN of 90%). However, crowdsourcing also involves manual work and hence comes at a higher cost. Combining the data from crowd and synthetic sources is a good compromise as it was seen to improve the performance (to 92% PPV). Synthetic data can be easily generated in abundance. The effect of changing the relative proportion of synthetic data in the training set can be seen from the results in the last two rows of Table 4. They suggest that increasing the proportion of synthetic data can boost the performance though the quantum of improvement appears small. The reasons for the same could be that the lesion masks were randomly chosen and hence the actual variability in data was not captured in training. A better scheme to ensure this variability could yield better improvement. Based on the encouraging results, future work can be directed towards exploring such solutions for other abnormality detection problems as well other modality images.
Acknowledgment
This work was supported by the Department of Electronics and Information Technology, Government of India under grant: DeitY/R&D/TDC/13(8)/2013.
Gulshan V. Peng L. Coram M. et al.: ‘Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs’, JAMA, 2016, 316, (22), pp. 2402–2410 (doi:
deBrebisson A. Montana G.: ‘Deep neural networks for anatomical brain segmentation’, CoRR, vol. abs/1502.02445, June 2015
Mitry D. Peto T. Hayat S. et al.: ‘Crowdsourcing as a novel technique for retinal fundus photography classification: analysis of images in the EPIC Norfolk cohort on behalf of the UKBiobank Eye and Vision Consortium’, PLoS ONE, 2013, 8, (8), pp. 1–7 (doi:
Maier‐Hein L. Mersmann S. Kondermann D. et al.: ‘Crowdsourcing for reference correspondence generation in endoscopic images’. Medical Image Computing and Computer‐Assisted Intervention (MICCAI), Boston, MA, USA, September 2014, pp. 349–356
Ganz M. Kondermann D. Andrulis J. et al.: ‘Crowdsourcing for error detection in cortical surface delineations’, Int. J. Comput. Assist. Radiol Surg., 2017, 12, pp. 12–161 (doi:
Maier‐Hein L. Mersmann S. Kondermann D. et al.: ‘Can masses of non‐experts train highly accurate image classifiers?’. Medical Image Computing and Computer‐Assisted Intervention (MICCAI), Boston, MA, USA, January 2014, pp. 438–445
Maier‐Hein L. Rob T. Grohl J. et al.: ‘Crowd‐algorithm collaboration for large‐scale endoscopic image annotation with confidence’. Medical Image Computing and Computer‐Assisted Intervention (MICCAI), Athens, Greece, October 2016, pp. 616–623
Albarqouni S. Baur C. Achilles F. et al.: ‘AggNet: deep learning from crowds for mitosis detection in breast cancer histology images’, IEEE Trans. Med. Imag., 2016, 35, pp. 1313–1321 (doi:
Collins L. Zijdenbos A. Kollokian V. et al.: ‘Design and construction of a realistic digital brain phantom’, IEEE Trans. Med. Imaging, 1998, 17, (3), pp. 463–468 (doi:
Bonaldi L. Menti E. Ballerini L. et al.: ‘Automatic generation of synthetic retinal fundus images: vascular network’. Simulation and Synthesis in Medical Imaging: SASHIMI, Held in Conjunction with MICCAI, Loughborough, UK, October 2016, pp. 167–176
Havaei M. Davy A. Warde‐Farley D. et al.: ‘Brain tumor segmentation with deep neural networks’, Med. Image Anal., 2017, 35, pp. 18–31 (doi:
Goodfellow I. Pouget‐Abadie J. Mirza M. et al.: ‘Generative adversarial nets’, Adv. Neural. Inf. Process. Syst., 2014, 27, pp. 2672–2680
Rezaei M. Harmuth K. Gierke W. et al.: ‘Conditional adversarial network for semantic segmentation of brain tumor’, CoRR, vol. abs/1708.05227, August 2017
Nie D. Trullo R. Petitjean C. et al.: ‘Medical image synthesis with context‐aware generative adversarial networks’. Medical Image Computing and Computer‐Assisted Intervention, MICCAI, Quebec City, Canada, September 2017, pp. 417–425
Costa P. Galdran A. Meyer M. et al.: ‘Adversarial synthesis of retinal images from vessel trees’. Image Analysis and Recognition: 14th Int. Conf., ICIAR, Montreal, Canada, 2017, pp. 516–523
Virdi T. Guibas J.T. Li P.S.: ‘Synthetic medical images from dual generative adversarial networks’, ArXiv e‐prints, September 2017
Shankaranarayana S.M. Ram K. Mitra K. et al.: ‘Joint optic disc and cup segmentation using fully convolutional and adversarial networks’. Fetal, Infant and Ophthalmic Medical Image Analysis: OMIA Held in Conjunction with MICCAI, Québec City, Canada, 2017, pp. 168–176
Joshi G. Sivaswamy J.: ‘Colour retinal image enhancement based on domain knowledge’. Indian Conf. on Computer Vision, Graphics and Image Processing, Bhubaneswar, India, December 2008, pp. 591–598
Kauppi T. Kalesnykiene V. Kamarainen J.K. et al.: ‘Diaretdb1 diabetic retinopathy database and evaluation protocol’, 2007
Decencière E. Zhang X. Cazuguel G. et al.: ‘Feedback on a publicly distributed database: the Messidor database’, Image Anal. Stereol., 2014, 33, pp. 231–234 (doi:
Wang Z. Conrad Bovik A. Rahim Sheikh H. et al.: ‘Image quality assessment: from error visibility to structural similarity’, IEEE Trans. Image Process., 2004, 13, (4), pp. 600–612 (doi:
Ronneberger O. Fischer P. Brox T.: ‘U‐net: convolutional networks for biomedical image segmentation’, CoRR, vol. abs/1505.04597, May 2015
Prentasic P. Loncaric S. Vatavuk Z. et al.: ‘Diabetic retinopathy image database (DRiDB): a new database for diabetic retinopathy screening programs research’. Int. Symp. on Image and Signal Processing and Analysis (ISPA), Trieste, Italy, 2013, pp. 704–709
Giancardo L. Meriaudeau F. Karnowski T. et al.: ‘Exudate‐based diabetic macular edema detection in fundus images using publicly available datasets’, Med. Image Anal., 2012, 16, pp. 216–226 (doi:
Maninis K.‐K. Pont‐Tuset J. Arbelaez P. et al.: ‘Deep retinal image understanding’. Medical Image Computing and Computer‐Assisted Intervention (MICCAI), Athens, Greece, 2016, pp. 140–148
Kohler T. Budai A. Kraus M. et al.: ‘Automatic no‐reference quality assessment for retinal fundus images using vessel segmentation’. Int. Symp. on Computer‐Based Medical Systems, CBMS, Porto, Portugal, 2013, pp. 95–100
Prentasic P. Loncaric S.: ‘Detection of exudates in fundus photographs using deep neural networks and anatomical landmark detection fusion’, Comput. Methods Programs Biomed., 2016, 137, pp. 281–292 (doi:
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2018. This work is published under http://creativecommons.org/licenses/by-nc/3.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Annotations are critical for machine learning and developing computer aided diagnosis (CAD) algorithms. Good performance of CAD is critical to their adoption, which generally rely on training with a wide variety of annotated data. However, a vast amount of medical data is either unlabeled or annotated only at the image‐level. This poses a problem for exploring data driven approaches like deep learning for CAD. In this paper, we propose a novel crowdsourcing and synthetic image generation for training deep neural net‐based lesion detection. The noisy nature of crowdsourced annotations is overcome by assigning a reliability factor for crowd subjects based on their performance and requiring region of interest markings from the crowd. A generative adversarial network‐based solution is proposed to generate synthetic images with lesions to control the overall severity level of the disease. We demonstrate the reliability of the crowdsourced annotations and synthetic images by presenting a solution for training the deep neural network (DNN) with data drawn from a heterogeneous mixture of annotations. Experimental results obtained for hard exudate detection from retinal images show that training with refined crowdsourced data/synthetic images is effective as detection performance in terms of sensitivity improves by 25%/27% over training with just expert‐markings.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 Center for Visual Information Technology, IIIT Hyderabad, Hyderabad, India