Introduction
Early detection of polyps in colonoscopies is an essential task for the prevention of colon-cancer deaths. Accurate polyp detection is essential for colonoscopy performance since every 1% increase in the polyp-detection rate can decrease the interval of colorectal cancer incidence by 3% []. However, the adenoma-detection rate obtained by colonoscopists varies from 7 to 53%, even though the adenoma-prevalence rate is estimated to be higher than 50% in the screening-age population []. The previous meta-analysis also showed that approximately 26% of neoplastic diminutive polyps was missed in a single colonoscopy []. Therefore, there is a demand for an image-based computer-assisted diagnosis (CAD) system for colonoscopies to achieve improved colonoscopy performance.
To build a high-performance CAD system, we need an accurate classifier that correctly discriminates polyp-appearing scenes from non-polyp scenes in a sequence of images captured by the colonoscopy. Such stable polyp-scene detection has been a challenge for many years []. Previous methods are based on two-dimensional (2D) image processing of each still image in a colonoscopic video. MICCAI 2018's endoscopic vision challenge (Endovis18) [] adopted image-sequence-based evaluation as a detection algorithm by using short colonoscopic videos. In this evaluation, many state-of-the-art methods produced imbalanced results with biased sensitivity and specificity, demonstrating the difficulty of stable polyp detection. Furthermore, in EndoVis18's test dataset, the ratio between polyp and non-polyp scenes was about 2:1. Their evaluation scheme was biased towards sensitivity. Therefore, if we applied the methods submitted to EndoVis18, we may obtain many false positives (FPs). Moreover, state-of-the-art methods [] are variants of object detection methods [] and segmentation methods []. Making a large dataset is a notably difficult task for these methods since the training step requires annotation labels for the target object regions in each image.
For stable polyp detection, temporal coherence in colonoscopy views is a key structural factor, as shown in Fig. . Spatiotemporal feature extraction via a three-dimensional convolutional neural network (3D CNN) [] utilises temporal coherence in successive colonoscopic images. In previous studies [], a 3D CNN method achieved about 0.80 sensitivity and specificity in an image-sequence evaluation. These results were attained by a model trained with short videos of only 102 polyps and early stopping.
[IMAGE OMITTED. SEE PDF]
For further improvement in the detection accuracy of a 3D CNN, we need a larger training dataset. Unlike well-known object-detection [] and segmentation methods [], assembling a large training dataset is possible for a 3D CNN since it requires only weak annotation, which represents the existence of a polyp in each frame without annotation of the polyp location. However, learning for a 3D CNN with large colonoscopic videos involves an imbalance. The ratio between the polyp and non-polyp scenes is quite imbalanced in colonoscopic videos since most scenes do not include a polyp. This property also leads to imbalanced classification results. Furthermore, a 3D CNN is prone to overfitting since it has larger parameters than the usual 2D CNNs.
This paper tackles the imbalance problem of large-scale datasets in deep learning for automated polyp-scene classification in colonoscopic videos. We propose a new residual learning procedure from an imbalanced large dataset. We then introduce post-processing to train models as a way to achieve stable polyp-scene classification.
Methods
There are roughly two approaches to managing imbalanced datasets in machine learning []: using the weighting loss function and manipulating datasets. Where the data size is small, new data is synthesised as data augmentation before the model is trained []. However, the synthesising approach is costly when pre-processing a large video dataset. Moreover, generative adversarial networks (GANs)-based data augmentation is premature for real applications for the training of GAN easily diverse and trained GAN outputs images of unstable quality []. In unstable output images, some look realistic, and others not. For large-scale machine learning, loss-function weighting in the learning step is a common technique []. However, as the epoch continues, the effects of the dataset's imbalance appear. We introduce an online random subsampling technique for large scale deep learning. This reduces both the computational cost and the effect of imbalance.
Class-weight balancing in loss function
For an input query , which is a tensor containing T-frames of RGB images, we train the model , which returns the scores of K predefined classes for . We introduce a probabilistic representation of the trained model as
Subsampling excepting hard-negative samples
The class-weight balancing approach partly reduces the effect of imbalanced training data with early stopping. However, as the epoch continues, learning with class-weight balancing leads to overfitting of the dominant category, which contains most of the population in the training data.
For large-epoch training, we design a new subsampling approach. In the two-category classification of positive and negative categories as in polyp-scene classification, we assume that negative samples are the dominant category. In this large set of negative samples, we obtain specific samples, which include difficult patterns for correct classification, as hard-negative samples. We use hard-negative samples as the default set and randomly select a subset of the residual negative samples at each epoch in deep learning so that the ratio of positive and total negative samples is 1:1. We assume that these hard-negative samples distribute around the decision boundary between the two categories. Therefore, this procedure selects an important pattern for training and utilises all patterns in the negative samples after long learning with a balanced ratio of the two-category population.
Architecture for polyp-scene classification
We adopt spatiotemporal feature extraction with deep learning for our polyp-scene classification. In our previous work, we used a 3D CNN, C3DNet, for the spatiotemporal feature extraction []. C3DNet is a natural extension of 2D CNNs for 3D data classification [], as shown in Fig. . However, C3DNet is prone to early convergence and overfitting. In our previous work [], we constructed a C3DNet model with early stopping and class-weight balancing to avoid overfitting of non-polyp scenes. Our C3DNet model achieved 0.80 sensitivity and specificity for the test data extracted from the videos, where only typical protruding polyps appear.
[IMAGE OMITTED. SEE PDF]
Residual learning is a state-of-the-art method to avoid overfitting and extract fine features for classification. Therefore, we adopt a 3D version of the residual network (3D ResNet) to avoid overfitting []. In addition to introducing residual units, we introduce a wide residual architecture (Wide ResNet) [] by extending from a 2D convolution to a 3D convolution. The Wide ResNet doubles the number of kernels of the usual ResNet at the downsampling step. We set the weight decay to be 0.0001 for all residual units in our architectures and selected different sizes, numbers, and strides of kernels in the convolution to optimise the architectures for polyp-scene classification. However, the number, size, and stride width of the convolutional kernel and pooling are basically inherited from the C3DNet of our previous works []. Table summarises the architecture of our 3D ResNets. Models II and IV are wider versions of models I and III, respectively. In these architectures, the number of parameters of the 3D CNN and 3D ResNet is about 28 million and 14 million, respectively. To the best of our knowledge, this is the first time, 3D ResNet has been applied to polyp-scene classification.
Table 1 Architect of our 3D ResNet for polyp detection
Layer name | Architecture | |||
I | II | III | IV | |
conv1 | , stride | , stride | ||
max pool, stride | max pool, stride | max pool, stride | max pool, stride | |
conv2_* | ||||
conv3_* | ||||
conv4_* | ||||
conv5_* | — | |||
conv6_* | — | — | — | |
global average pooling and softmax. |
Post-processing
In real colonoscopic scenes, inappropriate still images for diagnosis can be captured, such as images that are out of focus and/or blurred due to the colonoscope moving too fast. Moving the colonoscope too quickly also results in some still images in which the polyps are out of sight. Other examples include occlusion by the colon wall and bubbles that conceal polyps in the polyp scenes. When the colonoscope moves too quickly, these inappropriate still images can appear in the successive images in polyp scenes. These images are a kind of noise and lead to misclassification in 3D CNN. Therefore, we use post-processing for the trained models. We set criteria . If a trained model outputs a value larger than over N-successive samples in an input video, the trained model outputs a polyp-scene label. This post-processing reduces FPs and preserves true positives (TPs).
Experiments and results
Dataset
We constructed a new dataset to validate our proposed learning method. We collected a total of 1027 full-HD-resolution 30-frames-per-second (fps) videos of 951 patients captured by the CF-HQ290ZI (Olympus, Tokyo, Japan) during a daily colonoscopy at Showa University Northern Yokohama Hospital with IRB (institutional review board) approval. We annotated all of the videos’ frames via two-step procedures with the annotation tool ELAN []. In the first step, two trained support staff performed annotations for all of the frames of each video. In the second step, two expert colonoscopists checked the annotated videos provided by the support staff. The annotation labels were the polyp existence, gross anatomy, polyp size, and observation conditions for all frames in the videos. The gross anatomy consists of the types of polyps (Is, Ip, Isp and IIa) and the observation conditions are white light, narrow-band imaging, and staining modes. These gross anatomical annotations suggest that both protruding and flat type polyps exist in our dataset. Note that about 60% of polyps in our dataset are the flat type. The annotated videos are divided into two groups: polyp and non-polyp videos. In this evaluation, we used the polyp existence and observation condition labels. Note that we used only scenes captured under the white light condition.
We divided these videos into the training, validation, and test datasets without duplication of the patients among them. The two expert colonoscopists selected typical polyp examples for detection by the CAD system for the validation and test datasets. The training, validation, and test datasets include both protruding and flat polyps. We used the validation data to check for overfitting in training and to select the trained model for evaluation. We used the test data to evaluate the trained model, which was selected by checking the validation-data results. Note that there is no duplication in the divided data. Therefore, we evaluated our trained model with unseen images. Furthermore, we selected hard-negative videos. We are already aware of difficult scenes in colonoscopic videos for polyp-scene classification via our previous works []. In these difficult scenes, we observed specific colonoscope actions, such as moving too close to and stopping near the colon walls. When colonoscopists capture the shapes of polyp-like structures, they bring the colonoscope close to the colon walls and stop it. These actions in non-polyp scenes lead to FPs because the 3D CNN recognises the actions in videos. In addition to these actions, the appearance of the specula and bubbles sometimes leads to FPs. Therefore, we selected non-polyp videos that include specula and bubbles.
We then extracted 16-frame chunks from each dataset. This extraction allowed for an overlap of eight frames between two successive chunks. The chunk size was (frames × height × width × channels), which is the common size in 3D convolution-based spatiotemporal feature extraction []. Our previous works also used this size []. Table summarises the details of our dataset.
Table 2 Summary of our polyp detection dataset
Polyp videos | Non-polyp videos | |||||||
Train | Validation | Test | Total | Train | Validation | Test | Total | |
#patient | 745 | 24 | 24 | 793 | 154 | 12 | 13 | 158 |
#videos | 797 | 31 | 40 | 868 | 165 | 15 | 11 | 159 |
#polyp | 1945 | 31 | 41 | 2017 | 0 | 0 | 0 | 0 |
#chunks | 90,096 | 2274 | 2184 | 94,554 | 284,510 | 21,436 | 20,291 | 326,237 |
Training of models
We trained a 3D CNN and 3D ResNet for our polyp-scene classification. We used an NVIDIA Tesla V100 16 GB GPU and a Keras with a TensorFlow backend. For the C3DNet, we trained two models with balanced weighting and our random subsampling, respectively. For these two trainings, we set an Adam optimisation algorithm with a default learning rate of lr = 0.00001 for 30 epochs. For the 3D ResNet, we used our random subsampling set at lr = 0.000001 for 70 epochs. We selected the best-trained models for each architecture by checking the classification accuracy with the validation data. These base learning rates were decided via preliminary experiments by checking the learning curves and selecting base learning rates that moderately lose values of loss functions for both the training and validation data.
Chunk-level evaluation
Using the trained models of each architecture, we computed the sensitivity and specificity against the test data. Sensitivity and specificity are the ratios of TPs in the polyp and non-polyp scenes, respectively, of the test data. Fig. illustrates the evaluation results of the ResNets in a chunk level. We omitted the results of two C3DNets, trained with class-weight balancing and our subsampling, respectively, since both models had almost zero sensitivity. Note that the C3DNet trained with our random samples gives higher sensitivity than the C3DNet trained with class-weight balancing for evaluation against the training data, even though both models are overfitted to the training data.
[IMAGE OMITTED. SEE PDF]
Scene-level evaluation
We performed a scene-level evaluation of the trained ResNets using 21 short video clips of 50–60 seconds. We created these 21 video clips from 40 polyp videos of 41 polyps in the test data. Each clip includes polyp scenes of one polyp and the length of one polyp scene is from 2 to 10 seconds. These clips were cropped from videos of actual endoscopies. Therefore, these polyp scenes include inappropriate still images due to quick colonoscope movement with bubbles and specula. This allows for practical evaluation as such inappropriate images inhibit polyp detection. Fig. shows an example of these 21 short video clips.
[IMAGE OMITTED. SEE PDF]
We fed these 21 clips to the four trained models with post-processing, where we extracted 16-frame chunks with 4-frame overlap between successive chunks. For post-processing, we set decision criteria for all four trained models. We chose these criteria to compare the four models in the same setting. We set N in post-processing to be 2 and 3 for models I and III, and II and IV, respectively. This setting of N was decided from the validation data classification.
In this evaluation, we conclude that correct detection is performed if the trained model detects at least one chunk in a polyp scene of a short video clip. We set the detection accuracy to be the ratio of correctly detected polyps in the 21 short video clips. Fig. shows an example of this evaluation. In Figs. , a polyp is correctly detected since the trained models classified several chunks for an approximately four-second polyp scene, as marked by the red lines. In addition to detection accuracy, we computed the average length of FP scenes among these 21 clips. FP scenes are non-polyp scenes that are incorrectly classified as polyp scenes. Table summarises the detection accuracy and average length of the FP scenes.
[IMAGE OMITTED. SEE PDF]
Table 3 Results of the scene-level evaluation
Model | Detection accuracy | Average length of FP scenes, s |
I | 0.86 | 1.5 |
II | 0.95 | 6.0 |
III | 0.71 | 0.8 |
IV | 1.00 | 4.5 |
Discussion
In the results of the chunk-level evaluation shown in Fig. , our ResNets with proposed random sampling reduced the overfitting of the non-polyp scenes, even though the C3DNet resulted in overfitting of zero sensitivity. Furthermore, random sample preserving of hard-negative samples achieved a high specificity of about 0.99. In the results in Fig. , the sensitivities of the trained models appear insufficient. However, the polyp scenes in the test data include many inappropriate noise images for the polyp-scene classification since the test data is sampled from actual colonoscopic videos. Furthermore, the inputs of the models partly include non-polyp scenes at the beginning and the end of the polyp scenes because the inputs in the evaluation are sets of 16 frames with an 8-frame overlap in successive chunks. Therefore, the test data is a very difficult set of practical-diagnosis scenes.
In the results in Fig. , we observe FPs even though we obtained high specificity in Fig. . In the chunk-level evaluation, we extracted chunks of polyp and non-polyp scenes from the polyp and non-polyp videos, respectively. However, in the scene-level evaluation, we extracted successive chunks of polyp and non-polyp scenes from the polyp videos. Therefore, non-polyp scene chunks include colon wall surface textures similar to the polyp scene textures. These similar textures sometimes lead FPs.
The results of the scene-level evaluation demonstrate the validity of our post-processing. In the examples shown in Fig. , the post-processing reduced many FPs and retained TPs of polyp scenes. The bubble and blurry scenes shown in Fig. were rejected through our post-processing. There are FPs after post-processing in Figs. . Models II and IV have higher representation ability of pattern than models I and III because they have the wide residual architecture of models II and IV. Therefore, models II and IV provide more FP scenes that have textures similar to the polyp scenes.
As summarised in Table , model II achieved the highest detection accuracy in the scene-level evaluation. However, the average length of the FP scenes was 6 seconds in model II, which is about 10% of the short video clips. This implies that model II is unstable. Model I achieved the second highest detection accuracy with the second shortest average length of FP scenes. This average length accounts for only about 3% of the short video clips. In other words, model I achieved 0.86 sensitivity and 0.97 specificity in the scene-level evaluation, indicating that model I achieves stable polyp detection. We showed the detection example in Fig. to three expert colonoscopists, who reported that model I appears to be the most practical model for real colonoscopy use.
Conclusions
Towards practical polyp-scene classification in colonoscopic videos, we tackled the imbalance problem of large-scale datasets in deep learning by establishing a new residual learning procedure from an imbalanced large dataset. We then introduced post-processing to the trained models as a way to achieve stable polyp-scene classification. We validated our proposed residual learning procedure and post-processing through experiments with a large dataset of 1027 colonoscopic videos. Our proposed method established stable polyp-scene classification in the scene level of an actual colonoscopic video with a moderate sensitivity of 0.86 and high specificity of 0.97. We conclude that this stable polyp-scene classification is helpful in real colonoscopies as a practical CAD system.
Funding and declaration of interests
Parts of this research were supported by AMED (19hs0110006h0003), MEXT KAKENHI (26108006, 17H00867, 17K20099), and the JSPS Bilateral Joint Research Project. Conflict of interest: None declared.
Corley, D.A., Jansen, C.D., Marks, A.M., et al.,: ‘Adenoma detection rate and risk of colorectal cancer and death’, N. Engl. J. Med., 2014, 370, pp. 1298–1306 (doi:
Urban, G., Tripathi, P., Alkayali, T., et al.,: ‘Deep learning localizes and identifies polyps in real time with 96% accuracy in screening colonoscopy’, Gastroenterology, 2018, 155, (4), pp. 1069–1078.e8 (doi:
van Rijn, J.C., Reitsma, J.B., Stoker, J., et al.,: ‘Polyp miss rate determined by tandem colonoscopy: a systematic review’, Am. J. Gastroenterol., 2006, 101, pp. 343–350 (doi:
Bernal, J., Sánchez, J., Vilariño, F.: ‘Towards automatic polyp detection with a polyp appearance model’, Pattern Recognit., 2012, 45, (9), pp. 3166–3182 (doi:
Bernal, J., Tajkbaksh, N., Sánchez, F.J., et al.,: ‘Comparative validation of polyp detection methods in video colonoscopy: results from the MICCAI 2015 endoscopic vision challenge’, IEEE Trans. Med. Imaging, 2017, 36, (6), pp. 1231–1249 (doi:
Gastrointestinal Image ANAlysis (GAINA) sub‐challenge in Endoscopic vision challenge. Available at https://giana.grand‐challenge.org/, accessed 9 July 2019
Shin, Y., Qadir, H.A., Aabakken, L., et al.,: ‘Automatic colon polyp detection using region based deep CNN and post learning approaches’, IEEE Access, 2018, 6, pp. 40950–40962 (doi:
Mo, X., Tao, K., Wang, Q., et al.,: ‘An efficient approach for polyps detection in endoscopic videos based on faster R‐CNN’. Proc. Int. Conf. on Pattern Recognition (ICPR), Beijing, China, 2018, pp. 3929–3934
Guo, Y.B., Matuszewskia, B.J.: ‘GIANA polyp segmentation with fully convolutional dilation neural networks’. Proc. VISIGRAPP, Prague, Czech Republic, 2019
Girshick, R., Donahue, J., Darrell, T., et al.,: ‘Rich feature hierarchies for accurate object detection and semantic segmentation’. Proc. Computer Vision and Pattern Recognition (CVPR), Columbus, Ohio, USA, 2014, pp. 580–587
Ren, S., He, K., Girshick, R., et al.,: ‘Faster R‐CNN: towards real‐time object detection with region proposal networks’. Proc. Neural Information Processing Systems (NIPS), Montréal, Canada, 2015, pp. 91–99
Redmon, J., Farhadi, A.: ‘YOLO9000: better, faster, stronger’. Computer Vision and Pattern Recognition (CVPR), Honolulu, Hawaii, USA, 2017, pp. 6517–6525
Liu, W., Anguelov, D., Erhan, D., et al.,: ‘SSD: single shot multibox detector’. Proc. European Conf. on Computer Vision (ECCV), Amsterdam, The Netherlands, 2016, (LNCS, 9905) pp. 21–37
Ronneberger, O., Fischer, P., Brox, T.: ‘U‐net: convolutional networks for biomedical image segmentation’. Proc. Medical Image Computing and Computer‐Assisted Intervention (MICCAI), Munich, Germany, 2015, (LNCS, 9351) pp. 234–241
Badrinarayanan, V., Kendall, A., Cipolla, R.: ‘SegNet: a deep convolutional encoder–decoder architecture for image segmentation’, IEEE Trans. Pattern Anal. Mach. Intell., 2017, 39, pp. 2481–2495 (doi:
Tran, D., Bourdev, L., Fergus, R., et al.,: ‘Learning spatiotemporal features with 3D convolutional networks’. 2015 IEEE Int. Conf. on Computer Vision (ICCV), Santiago, Chile, 2015
Hara, H., Kataoka, H., Satoh, Y.: ‘Learning spatio‐temporal features with 3D residual networks for action recognition’. Proc. Int. Conf. on Computer Vision Workshop (ICCVW), Venice, Italy, 2017, pp. 3154–3160
Itoh, H., Roth, H.R., Oda, M., et al.,: ‘Towards automated colonoscopy diagnosis: binary polyp size estimation via unsupervised depth learning’. Proc. Medical Image Computing and Computer‐Assisted Intervention (MICCAI), Granada, Spain, 2018 (LNCS, 11071), pp. 611–619
Misawa, M., Kudo, S., Mori, Y., et al.,: ‘Artificial intelligence‐assisted polyp detection for colonoscopy: initial experience’, Gastroenterology, 2018, 154, (8), pp. 2027–2029.e3 (doi:
Japkowcz, N.: ‘Learning from imbalanced data sets: a comparison of various strategies’. Proc. Int. Workshop on Learning from Imbalanced Data Sets, Austin, Texas, USA, 2000
Tran, T., Pham, T., Carneiro, G., et al.,: ‘A Bayesian data augmentation approach for learning deep models’. Proc. Neural Information Processing Systems (NIPS), Long Beach, California, USA, 2017, pp. 2797–2806
Ratner, A.J., Ehrenberg, H., Hussain, Z., et al.,: ‘Learning to compose domain‐specific transformations for data augmentation’. Proc. Neural Information Processing Systems (NIPS), Long Beach, California, USA, 2017, pp. 3236–3246
Miyato, T., Kataoka, T., Koyama, M., et al.,: ‘Spectral normalization for generative adversarial networks’. Proc. Int. Conf. on Learning Representations (ICLR), Vancouver, BC, Canada, 2018
Neyshabur, B., Bhojanapalli, S., Chakrabarti, A.: ‘Stabilizing GAN training with multiple random projections’. Proc. Int. Conf. on Learning Representations (ICLR), Vancouver, BC, Canada, 2018
King, G., Zeng, L.: ‘Logistic regression in rare events data’, Polit. Anal., 2001, 9, pp. 137–163 (doi:
sklearn.utils.class_weight.compute_class_weight, Documentatoin of scikit learn. Available at https://scikit‐learn.org/stable/modules/generated/sklearn.utils.class_weight.compute_class_weight.html, accessed 9 July 2019
Zagoruyko, S., Komodakis, N.: ‘Wide residual networks’. Proc. British Machine Vision Conf. (BMVC), York, UK, 2016
Wittenburg, P., Brugman, H., Russel, A., et al.,: ‘ELAN: a professional framework for multimodality research’. Proc. Fifth Int. Conf. on Language Resources and Evaluation, Genoa, Italy, 2006. Available at https://tla.mpi.nl/tools/tla‐tools/elan/
Hara, H., Kataoka, H., Satoh, Y.: ‘Can spatiotemporal 3D CNNs retrace the history of 2D CNNs and ImageNet?’. Proc. Computer Vision and Pattern Recognition (CVPR), Salt Lake City, Utah, USA, 2018, pp. 6546–6555
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2019. This work is published under http://creativecommons.org/licenses/by/3.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
This Letter presents a stable polyp‐scene classification method with low false positive (FP) detection. Precise automated polyp detection during colonoscopies is essential for preventing colon‐cancer deaths. There is, therefore, a demand for a computer‐assisted diagnosis (CAD) system for colonoscopies to assist colonoscopists. A high‐performance CAD system with spatiotemporal feature extraction via a three‐dimensional convolutional neural network (3D CNN) with a limited dataset achieved about 80% detection accuracy in actual colonoscopic videos. Consequently, further improvement of a 3D CNN with larger training data is feasible. However, the ratio between polyp and non‐polyp scenes is quite imbalanced in a large colonoscopic video dataset. This imbalance leads to unstable polyp detection. To circumvent this, the authors propose an efficient and balanced learning technique for deep residual learning. The authors’ method randomly selects a subset of non‐polyp scenes whose number is the same number of still images of polyp scenes at the beginning of each epoch of learning. Furthermore, they introduce post‐processing for stable polyp‐scene classification. This post‐processing reduces the FPs that occur in the practical application of polyp‐scene classification. They evaluate several residual networks with a large polyp‐detection dataset consisting of 1027 colonoscopic videos. In the scene‐level evaluation, their proposed method achieves stable polyp‐scene classification with 0.86 sensitivity and 0.97 specificity.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Graduate School of Informatics, Nagoya University, Nagoya, Japan
2 Digestive Disease Center, Showa University Northern Yokohama Hospital, Tsuzuki‐ku, Yokohama, Japan
3 Research Center for Medical Bigdata, National Institute of Informatics, Chiyoda‐ku, Tokyo, Japan