1. Introduction
Oral potentially malignant disorders (OPMDs) are one of the severe health issues across the globe [1]. OPMDs include oral leukoplakia [2], oral erythroplakia [3,4], oral submucous fibrosis, oral lichen planus [5], and oral lichenoid reaction [6], Other oral mucosal lesions (OML) include non-specific ulcerations, erythematous lesions, abscesses, median rhomboid glossitis, frictional keratosis, and pyogenic granuloma. The detection and diagnosis of these OPMDs as early as possible are essential for the dental specialist, as these anomalies might transform into oral cancer [7,8,9,10]. At the advanced stage, it is more challenging to cure oral cancer. The accepted method for the detection of OML and OPMDs is the conventional oral examination (COE) (visual inspection) [11,12]. The COE is exceedingly sensitive in detecting vicissitudes in the oral cavity because of the easy visibility of various oral cavity structures. However, the predictions of which COE-identified lesions will progress to oral cancer is dependent on the histopathological findings of epithelial dysplasia, as reported by an oral pathologist in biopsy tissue [13]. Currently, along with the use of the COE, clinicians/dental specialists use different additional examinations, e.g., the toluidine blue test (TB) [14], ViziLite® plus [15], VELscope® as shown in Figure 1, and Identafi® [16] as an adjunct to the COE in locating the area/areas to increase the predictive value for epithelial dysplasia. The histopathological diagnosis of epithelial dysplasias is currently the gold standard, as they have a higher potential for malignant change as compared to COE-identified OPMDs [17].
With the aid of these added tools, the precision in detecting areas that predict epithelial dysplasia in these oral lesions may improve. Although these adjunct tools may help the clinicians/dental specialists in detecting potential areas with epithelial dysplasia, the examination of the images relies on clinicians′ experience to discriminate amongst the potentially standard and anomalous regions [17,18]. The autofluorescent imaging device (Identafi, VELscope) permits the dentist to record video frames that are used for more detailed examinations.
Texture analysis is one of the crucial aspects of the vision system for differentiating between surfaces and objects. A surface is usually defined as rough (having a massive difference between high and low points) or silky (having little difference between low and high points), and a bumpy surface refers to touch [19]. For three decades, in digital image processing, texture has referred to visual configurations or the spatial organization of pel that regional intensity or color cannot explain adequately. Texture provides a better explanation of the content of the image structure as compared to an intensity illustrator, e.g., the average gray level, minimum/maximum gray level, or histogram of the local region. The texture appears to be a significant feature for an automated and semi-automated explanation of digital images for detecting the region of interest (ROI). It has a history of more than three decades in the field of educational, biomedical [20,21], military [22], commercial [23], web, and satellite image classification [24].
Medical image processing helps the clinicians by analyzing, in an objective way, the various kinds of images and different textures associated. As a result, the disease areas are identified more accurately from the standard ones. In this paper, the statistical behavior of the texture that occurs in the autofluorescence images is discussed to identify anomalous areas inside the oral cavity based on autofluorescence imaging [25]. In the case of biomedical images, textures occur randomly in the images but possess specific characteristics due to the symmetry in each body part of a living organism. This study proposes the use of the gray-level co-occurrence matrix (GLCM) parameters as features to identify the anomalous areas. It has to be mentioned that, to our knowledge, no image processing techniques based on textural pattern analysis have been attempted to date on the autofluorescence images from clinicians’ examinations using a machine learning approach. The main contribution of this article is as follows:
The proposed Healthcare Professional in the loop (HPIL) model acts as an aided tool for periodontists by automatically analyzing the VELscope® image of an oral cavity to find the ROI more precisely.
A texture-based machine learning algorithm using VELscope® images to discriminate OPMD and OML regions from a normal oral cavity.
The design of a Graphical User Interface (GUI) to assist clinicians in the classification of OPMDs.
The remainder of the paper is organized as follows: Section 2 describes the background of the algorithms used in this paper; Section 3 details the approach proposed, dataset information, and statistical testing criteria for result verification. Section 4 and Section 5 present the results and discussion, respectively, and Section 6 concludes the paper.
2. Background
2.1. State-of-the-Art Techniques and Necessity of Research
Screening for oral cavities implies searching for OPMDs and OML, typically before symptoms occur. Traditionally, the following steps have been followed by clinicians for screening and diagnosis inside the oral cavity. 1. Determining the background history of existing disease (if any): (a) the beginning, place, strength, occurrence, and period; (b) any irritation or discharges; (c) whether the disease has improved, remained constant, or worsened over the period. 2. Medical and drugs history (if any): (a) medical circumstances; (b) medications and antipathies; (c) tobacco and alcohol history (nature and time). 3. Medical examination: (a) extra mouth-cavity screening; (b) an intraoral check-up; (c) lesion screening using adjunctive visual tools such as direct fluorescence. 4. Differential analysis. 5. Diagnostic tests: (a) a biopsy (if needed). 6. Definitive diagnosis. 7. Suggested management [26]. With the advent of various diagnostic modalities, today, clinicians are using numerous approaches for the early diagnosis of OPMDs and OML [27]. ViziLite: ViziLite is an adjunctive tool working on the principle of tissue reflectance that has been used for the screening of the oral mucosa for “‘acetowhite” premalignant and malignant lesions. In recent times, a tissue reflectance-based screening device has been used for the oral cavity examination and is currently available as ViziLite [28]. Identafi®: The deep-penetrating multispectral lighting (three wavelengths) of the Identafi® enhances the early detection of abnormalities inside the oral cavity that might be cancerous. It uses the fluorescence and reflectance principles to highlight the properties of the epithelial cells. Hence, the system demonstrates an enhanced visualization of mucosal abnormalities, i.e., mouth cancer or premalignant cancerous stages that may not be visible to the human eye [29]. VELscope®: VELscope® stands for Visually Enhanced Lesion scope [30]. By lesion, the vendor means any abnormality in the tissue of an organism. It works on the principle of the direct visualization of tissue fluorescence, and the variations in fluorescence intensities help the clinicians to differentiate the lesion regions from the normal area of the oral cavity. VELscope® is used as an oral-cancer-screening device. Today, it is a supplementary oral examination device for non-invasive screening. With the help of VELscope, clinicians cannot determine whether the lesions are cancerous or not. Actually, it eases the detection and location of abnormal tissues that are not visible to the naked eye. When the oral cavity of a human is exposed to near ultraviolet (NUV) light, the normal cells will glow (fluoresce) brighter; on the other hand, abnormal tissues, whether cancerous or precancerous, will absorb the fluorescent light and appear darker (black). The light-absorbing property of abnormal and light-reflecting property of normal tissues of the oral cavity allow the clinicians to directly visualize the difference between the two tissues. However, a biopsy is still needed for the diagnosis of the specific detected disease.
The working principle of VELscope [31] comprises a handle-top light source attached to a holding unit for imaging. The illuminating source uses a “120 W metal halide arc (MH)” spotlight with a fundamental elongated mirror optimized for NUV (azure reflection). The flexible pinhole wheel is used to change the power attached to the light conductor. The light is attached to the light source via a holding unit with a “0.59 Numerical Aperture (NA)”. The handheld unit projects excitation rays onto the mouth cavity and a coaxial perceiving port with illumination for “fluorescence visualization (FV)”. A two-piece lens system, f = 25 mm, gathers light and transfers it to the soft muscle through a “low fluorescence excitation filter” (EX). A “dichroic mirror (DM)” is responsible for coaxial excitation and visualization paths. An “emission filter (EM)” allows the green–red glowing light to pass and stops the blue light; however, a “notch filter (NF)” splits the fluorescent light range into green and red colors.
All these devices depend on human visualization that is followed by a biopsy for the validation of the results. The effectiveness of all these devices is inconsistent [32,33]. The literature shows that with the help of these devices, currently, there is no noticeable improvement in the detection of OPMDs from the standard COE routinely performed by clinicians. These devices (VELscope®, Identafi) also help us to capture images of the oral cavity, but today, all the images captured via these devices are used for documentation proposes by clinicians; on the other hand, it is rather difficult for the clinicians to obtain images of the oral cavity with the Identafi. In the case of the VELscope®, we can mount the camera on the device; clinicians can examine the oral cavity using the camera screen and also capture the images. In order to examine these images to detect abnormalities of the oral cavity, biomedical image processing plays an important role in the diagnoses of many other cancerous diseases in different areas of the human body.
In recent years, a lot of research has been performed on the detection and identification of abnormalities inside the oral cavity using image-processing techniques [34,35,36,37]. M. Muthu Rama Krishnan et al. [38] used 114 images of oral cavities, of which 67 images were normal and the remaining were from patients suffering from oral sub-mucous fibrosis without dysplasia. All the images were captured with a “Carl Zeiss Microscope” using “H&E stained histological sections” under a “10x objective (NA 0.25)”. At a resolution of 1.33 µm and a dot size of 0.63 µm, these images were then pre-processed by using median and histogram techniques to enhance the details in the images, which was followed by the fuzzy logic technique. Later on, textural analysis was performed using wavelets and Gabor wavelets. The calculated features were than selected via Kullback–Leibler (KL). These selected features were then passed to a Bayesian classifier and Support Vector Machine (SVM) for the screening and classification of oral sub-mucous fibrosis (OSF). The results show that the SVM with the linear kernel function provides an improved classification accuracy of 92%, in comparison with the Bayesian accuracy of 76.83%.
Tze-Ta Huang et al. [39] used VELscope® images of oral cancer or precancerous lesions and a control group with normal oral mucosa patients. The abnormalities in the images were chosen as the ROIs. The average intensity and heterogeneity of the ROI were calculated. A quadratic discriminant analysis (QDA) was employed to compute boundaries based on sensitivity and specificity. This differentiated the normal oral mucosae from precancerous/oral cancer lesions with a specificity of 92.3% and a sensitivity of 97.0%.
Anuradha et al. [40] proposed statistical feature extraction to classify oral cancers. They used 27 dental X-ray images to test their algorithm. The first step of pre-processing was performed using image enhancement; at the second stage, image segmentation was performed with the help of Marker Controlled Watershed segmentation, which was followed by a feature extraction method using the GLCM. These features were then passed to the SVM classifier to classify the cancer. The accuracy obtained with the help of the proposed system was 92.5%.
From the literature, it is apparent that most of the work has been performed on oral cancer diagnosis using complicated and invasive procedures such as spectroscopy and biopsies. The research gap in the classification of OPMDs and OML using machine learning techniques based on VELscope image analysis has not been addressed yet. Previous work has shown more-quantitative analysis of the intensity and heterogeneity of VELscope® autofluorescence images to discriminate between cancer and precancerous cells. There is no such computer-based textural analysis system to help clinicians to find ROIs more efficiently and effectively.
2.2. Quadtree
A quadtree is a partition of the image in which successively deeper levels represent more pleasing subdivisions of image regions, as shown in Figure 2, in which Level 0 represents the full image and Level 1 divides the above region into four equally sized regions, and this process continues until the required size of the image region is achieved. The obtained sub-image regions are called super-pixels. One of the limitations of the quadtree is that, at each level of subdivision, the size of the image must be an even number. For images that allow multiple subdivisions, a trade-off between the computation time and the resolution of the super-pixels is required. All the original images in our database have a size equal to 1792 × 1792. A quadtree division was performed to produce a 28 × 28 super-pixel matrix, where each super-pixel was of the size 64 × 64. For example, for one image, if a super-pixel size of 32 × 32 is chosen, the number of super-pixels is multiplied by 4, resulting in 6912 × 4 = 27,648 co-occurrence matrices to be computed. Hence, choosing a 64 × 64 super-pixel size ensures a reasonable computation time as well as resolution for clinicians to perform biopsies of OPMDs accurately.
2.3. GLCM Texture Approach
The GLCM is one of the most used textual analysis techniques that involves the statistical sampling of the gray levels that occur concerning another gray level in an image. Haralick [41,42] was the first to explain the principle of the co-occurrence probability for extracting different features in images. The GLCM works on the principle of how different combinations of pixels occur in an image or local region. In our research, fourteen parameters were used to differentiate between anomalous and standard regions.
The GLCM has drawn significant interest in recent research performed on texture analysis Hossain and Parekh [43] used a textural analysis technique for color information to recognize texture with different channel combinations, e.g., gb, rr, bb, rg, gg, rb, gr, br, and bg. Neha and Amandeep [44] used texture analysis using the GLCM to differentiate different regions (mountains, rocks, and rivers) from satellite images. Nitish and Vrushsen [45] worked on textural feature analysis for four different categories of brain tumors and classified them into different categories on the basis of different GLCM parameters. Jamesa and Dasarathy [46] performed segmentation using statistical patterns on fused biomedical images taken from different sensors, at different times, and from different viewpoints. They fused two different types of cancer and tumor images: CT and MRI images. M. Reza et al. [47] used 11,000 medical X-ray images to perform a texture analysis comparison of different types of techniques such as the local binary pattern and Canny edge operator.
3. Materials and Methods
3.1. Proposed System
The proposed Healthcare Professional in the loop (HPIL) model acts as an aided tool for periodontists by automatically analyzing the VELscope® image of an oral cavity to find the ROI more precisely. The state-of-the-textural-art model assists the periodontist with a more accurate diagnosis that results in reducing the false-positive and false-negative cases. Figure 3 depicts the overall workflow of the proposed HPIL model in comparison with the previous, existing screening procedure for the diagnosis of oral cavity abnormalities.
3.2. Healthcare Professional in the Loop (HPIL)
A number of authors have proposed analyzing medical images based on texture statistics or other features such as wavelets, the Gabor filter, and the fractal dimension. In this paper, the proposed method for the classification of OML and OPMDs is based on the division of images into sub-regions using a quadtree. Hence, super-pixels are created and analyzed with a GLCM textural analysis technique.
Figure 4 shows the overall textural analysis of VELscope® images to classify the OPML and distinct regions. As the first step, an input RGB autofluorescence image is converted into a grayscale image. A pre-processing step is needed in the case of the presence of the device area in the image. The circular Hough transform (CHT) is used to detect the presence of the device part in the image. As a result, either the same grayscale image (if no device is detected) or a reduced image containing the ROI (i.e., the circular part of the image containing the inspected oral cavity area) is produced. A detailed description of CHT is provided in the “Experimental Results” section. Then, the image is subdivided into a certain number of super-pixels using the quadtree method. As the dimensions of these images are in the power of 2, it is easier to divide them into super-pixels all having the same size. Thus, the super-pixels are analyzed with the help of the GLCM to extract several features for each super-pixel. A feature selection based on linear discriminant analysis (LDA) is then performed in order to rank the GLCM features and enhance the later classification. Finally, the ranked features are classified using a k-NN classifier to identify the standard and anomalous regions (Figure 4). The performance of the algorithm was evaluated with the help of the ground truth provided by dental specialists of standard and anomalous regions of the oral cavity and expressed as the sensitivity, specificity, and accuracy.
3.3. Dataset of Auto-Fluorescence Images
The autofluorescence images were captured with a Cannon D620 using a VELscope® device by clinicians/dental specialists. The experiments were conducted on the autofluorescence images captured by clinicians/dental specialists from the Oral Cancer Research Co-ordination Centre (OCRCC), Faculty of Dentistry, at the University of Malaya (UM). This study was approved by the Medical Ethics Committee of the Faculty of Dentistry, University of Malaya (MEC: OI DF1501/0088(L)). Twenty-two patients were involved in this experimental analysis. Out of 30, 8 subjects in our database had more than one image of the oral cavity (e.g., tongue (lift-up), lingual frenum, etc.) due to the existence of oral mucosal lesions (OML) or oral potentially malignant disorders (OPMDs) in different oral cavity regions. The images of 15 patients were captured with a 2nd generation VELscope®, and the remaining images were captured with a 3rd generation VELscope®. All the patients were suffering from a certain kind of OPMD or precancerous stage that was confirmed via histopathology. The standard and anomalous regions were appropriately defined by the clinicians, which was important and useful for performing this analysis. Twenty-two autofluorescent images were used in this study involving 24 lesions. The clinical and histopathological diagnoses of all these lesions are provided in Table 1. The ground truth for each patient image was prepared with the help of expert dental specialists (AR). The clinicians correctly delineated the standard and anomalous regions in copied images.
3.4. Data Acquisition Using VELscope®
To record the VELscope® images of the oral cavity region [48,49], explicit camera settings needed to be set to obtain focused and clear images of the oral cavity.
-
The first significant stage is selecting the camera as there are a number of digital cameras accessible in the shop to be adapted for use with the VELscope® device, such as the Canon A620 and G7, and Nikon P5000. These single monocle reflex DSLR can be directly attached to the VELscope®.
-
We adopted the “Canon A620” to acquire the images using the VELscope®. Table 2 depicts the setup configurations that research studies have used to capture VELscope® images of the oral cavity. Figure 5 and Figure 6 show the effects of the oral cavity images with/without the correct settings for the “Canon A620”, respectively.
3.5. GLCM
In this paper, the GLCM was used to find the textural pattern of the standard and anomalous regions of the oral cavity. In this research study, we evaluated the 10 GLCM features (Angular Second Moment, Contrast, Correlation; Sum of Squares: Variance, Inverse Difference Moment, Sum Average, Sum Variance, Sum Entropy, Entropy, and Difference Variance) [41,42]; with the help of these parameters, we could perform the statistical texture analysis of the different regions of the image. These ten different statistical parameters (f1, f2, f3….. f10) are given below.
Whereas is the entry in the standardized gray-tone spatial-dependence matrix, is the entry in the marginal probability matrix, . is the number of distinct gray levels in the quantized image. Similarly,, k = 2,3,….., , and k = 2,3,….., [50,51,52]
Angular Second Moment (f1) (the Angular Second Moment is also known as the Uniformity or Energy):
Contrast (f2) (processes the local variations in the gray-level co-occurrence matrix):
Correlation (f3) (measures the joint probability of occurrence of the specified pixel pairs):
where , ,, and are the means and standard deviations of and .Variance (f4) (sum of squares):
Inverse Difference Moment (f5) (the Inverse Difference Moment is the confined homogeneity; it is high when the local gray level is uniform and vice versa):
Sum Average (f6):
Sum Variance (f7):
Sum Entropy (f8):
Entropy (f9) (provides the sum of squared elements in the GLCM, also known as the Uniformity or the Angular Second Moment):
Difference Variance (f10):
3.6. Feature Selection Based on Linear Discriminant Analysis (LDA)
The main aim of using LDA was to differentiate the set of features that provided the best discrimination between the standard and anomalous regions of the oral cavity. LDA is a mathematically robust technique and is often used to produce models that have accuracies that are as good as those of more complex methods. It is one of the powerful techniques used in classification, statistical analysis, and pattern recognition to find the best linear combination of features. It was initially developed in 1936 [53,54]. The basic working principle of LDA works on the concept of searching for linear combinations of variables (predictors) that help to separate the classes (targets). To separate the two regions, Fisher defines the scoring function as shown in Equation (1), in which and are the means of the particular regions; and are the coefficients that need to be calculated to obtain the value.
(1)
(2)
With , the Score Function to be maximized, as shown in Equation (2), C is the covariance matrix, is the coefficient matrix (linear model coefficient), and is the mean of the different classes; i.e., and are the represented means of two different classes, respectively.
Equation (3) can be represented in the general form of a scoring function:
(3)
Meanwhile, the linear coefficient matrix for ( and ) can be calculated:
(4)
The covariance matrix (C) can be calculated with the following mathematical Equation (5):
(5)
represent the numbers of observations (values) in the first and second classes, respectively.
Once the coefficient and linear combination are found, the effectiveness of the discriminative vector is calculated—the Mahalanobis distance28—to calculate the difference between two regions for particular features, as shown in Equation (6).
(6)
= the Mahalanobis distance between two regions.
When the resultant value is enormous, it means there is a small overlap between the two classes. From the LDA ranking of the features, the top features are selected to perform the classification.
3.7. K-Nearest Neighbors (KNN) Classifier
In the early 1970s, k-nearest neighbors (KNN) was used in statistical measure analysis and the recognition of patterns [55,56,57]. It is one of the simplest and most important non-parameter algorithms that saves all the training images and classifies new test images on the basis of similarity measure analysis. One of the main advantages of using KNN is that the classification rules are produced by the case samples (classes) without the additional need for any other parameter or data.
The KNN algorithm predicts the test case category based on the training set and tries to find the most similar K number (of the nearest neighbor) of the test case. One of the simplest ways to decide for the test case is to calculate the distance of each sample and find the smallest path to predict the class of a test case.
3.8. Classifier Performance
The efficiency of the KNN classifier for each combination of textural features was analyzed via three statistical measures, i.e., the accuracy, sensitivity, and specificity. The accuracy represents the actual results among all the numbers of regions analyzed as the system divides an image into sub-images of window 64 × 64. These sub-images are considered as anomalous regions if all their values lie in the red region, as shown in the ground truth in Figure 7. On the other hand, if a sub-image contains standard and anomalous regions, it is considered as a standard sub-image. The reason for setting this rule is to obtain an entire anomalous area for biopsy.
3.9. Evaluation Criteria
The evaluation was performed based on these three statistical parameters (accuracy, sensitivity, and specificity) as shown in Equations (7)–(9), respectively.
(7)
(8)
(9)
True Positive (TP): the anomalous disease area is properly identified by the classifier.
False Positive (FP): the standard region is identified as an anomalous region by the classifier.
True Negative (TN): the standard region is adequately identified by the classifier.
False Negative (FN): the anomalous region is identified as a standard region by the classifier.
4. Experimental Results
Figure 8 shows an RGB image of the lateral tongue having mild dysplasia, but the disease area is not properly clear. Figure 9 shown a VELscope® image in which the disease area appears as black, but the accurate location of the disease area is identified by the clinician, as shown in Figure 7. It is observed that the VELscope® device circular region appeared in a number of images; this noisy region might affect classification accuracy. Therefore, the Hough transform was adopted as a pre-processing algorithm to remove the circular region from the VELscope® images.
4.1. Pre-Processing (Edge Detection and CHT)
To remove the noise from the VELscope® image, a pre-processing step was performed. The pre-processing was divided into two parts. First, the edges were enhanced using the Deriche–Canny detector [58], as shown in Figure 10. Later on, the CHT was applied to the edge-enhanced images. In the selected database, research has shown two cases; in the first case, the device area was present, which has been removed by CHT, and in the second case, the device area is absent.
4.2. Case 1: When the Device Area is Present
Figure 11 shows a grayscale VELscope® image in which the camera region appears. Initially, the grayscale VELscope® images were used to detect the edges, especially the circular VELscope® region, using the Deriche–Canny edge detector [59], as shown in Figure 12. The edge-detected pre-processed image was then passed to the circular Hough transform (CHT). In Figure 13, the detected circular VELscope® region is shown and highlighted with a yellow-color circle; the detected region without the circular region was extracted, as shown in Figure 14.
4.3. Case 2: When the Device Area is Absent
Figure 15 shows a VELscope® image with no device area present. Initially, the RGB VELscope® image was converted into grayscale, as shown in Figure 16; the grayscale VELscope® images were used to detect the edges, especially the circular VELscope® region, using the Deriche–Canny edge detector as shown in Figure 17. The edge-detected pre-processed image was then subjected to CHT, and CHT tried to find the circular VELscope® region in an image. Figure 18 shows that no circle was detected; it depicts that the device area was absent.
After this pre-processing step, the resultant image was passed to the quadtree algorithm to divide the image into small windows. Once the image was divided into super-pixels, the GLCM was applied to each super-pixel to calculate the features; a co-occurrence matrix size of 256 × 256 was calculated at 0°. This co-occurrence size for the matrix helped us in a more detailed textural analysis of each pattern in an image without rounding off values from the grayscale super-pixel. If the co-occurrence matrix size had been less than 256 × 256, then after rounding off the grayscale values to the nearest values, a lot of textural information related to the ROI might have been lost. As observed by the clinicians (project collaborators), we also know that in comparison to the standard regions, there is a fractional variation in the textural regions of OMPD. Therefore, selecting the co-occurrence size of 256 × 256 helped us to differentiate them properly, so the textural analysis technique GLCM was applied to each window of an image and 10 features were calculated, i.e., the variance, correlation, difference entropy, sum variance, sum entropy, entropy, sum average, measure of co-relation, inverse difference moment, and contrast. Figure 19 depicts the textural features obtained via the GLCM using VELscope® images.
Once all the features were calculated, the classification of the anomalous and standard regions was performed using the KNN classifier using all 10 GLCM features; however, the classification accuracy was less than 50%; in order to improve the classification accuracy, a feature selection process was performed using linear discernment analysis (LDA). Using LDA, features were arranged in descending order, as shown in Table 3, based on their capabilities of differentiating standard and anomalous regions. If the values of the features were greater, it meant that a particular feature was suitable for separate the anomalous and the standard region. In Figure 20, a graphical analysis of each feature selected via LDA using probability density phenomena is shown. These graphical results confirm that the features selected via LDA distributed the values among two regions (standard and anomalous), which would be helpful for classification. These features were then used alone as well as in combinations to find the highest accuracy with the KNN classifier, as shown in Table 3. The first eight GLCM features (the variance, correlation, inverse difference moment, sum average, sum variance, sum entropy, entropy, and difference entropy) selected using LDA provided the highest accuracy. However, the information measure of correlation and contrast features depict that the disease region and standard region overlapped with each other; these features would impact the overall system accuracy if the system used them for classification.
4.4. Classification
Before starting the classification using KNN, the images in the database were divided into six groups. Each group contained five images. The proposed research study performed five-fold cross validation, by training the classifier on four images, testing it on one image, and repeating this process five times for each group so that all the images passed through the testing phase.
Following the ranking obtained from LDA and starting with the first parameter, at each stage, a parameter was added, and the performance was calculated using the KNN classifier. The results are shown in Table 4. The performance increased until reaching Parameter 8. From Parameter 9, the performance decreased—Figure 21 shows the graphical analysis of the parameters with the statistical results.
5. Discussion
Other than our proposed textural model, we evaluated the robustness and efficiency of our HPIL model in comparison with other existing texture descriptors using our VELscope® database, such as gradient directional pattern (GDP) [60], gradient directional pattern 2 (GDP2), geometric local textural patterns (GLTP) [61], improved Weber local descriptor (IWLD) [62], localized angular phase (LAP) [63], local binary pattern (LBP) [64], local directional pattern (LDIP) [65], local directional pattern variance (LDiPv), inverse difference moment standardized (IDN) [66], local directional number pattern (LDNP) [67], local gradient increasing pattern (LGIP) [68], local gradient patterns (LGP) [69], local phase quantization (LPQ) [70], local ternary pattern (LTeP) [71], local tetra pattern (LTrP) [72], monogenic binary coding (MBC) [73], local frequency descriptor (LFD) [74], and local mapped pattern (LMP) [75]; however, the overall results with these textural descriptors were quite modest (as shown in Table 5) as compared to those obtained with our HPIL model, which makes it not suitable for classifying OPMD and standard regions.
Table 6 depicts the comparison of the proposed approach with the existing work. In previous research, the analysis and screening of VELscope® images demanded a conventional oral examination (COE) followed by a biopsy. The existing system still relies on periodontist expertise for the annotation of the ROI to perform a biopsy. The sensitivity of the diagnosis of oral pathology ranges from 15.3% to 76%. On the contrary, the proposed textural analysis algorithm automatically analyzes the VELscope® images and detects the ROI more precisely. The system achieved a more reliable sensitivity and specificity of 85 ± 5% and 84 ± 3%, respectively. Our proposed model acts as an aided tool for the periodontist to detect OPMD and precancerous cells precisely, which will result in a significant decrease in the clinician’s workload and reduce the screening time by analyzing the VELscope® image automatically. Today, VELscope® images are solely analyzed by the periodontist. The screening of these images demands high professional proficiency to find the exact location of OPMD or precancerous cells very precisely so that it can be used for biopsy. In addition, the labeling of VELscope® images is a troublesome task for periodontists due to its tiring work routine. Currently, there is no such existing algorithm available for analyzing VELscope® images to identify OPMD/OML regions automatically and help periodontists in diagnosis via biopsy.
Figure 22 shows the comparative ROC analytics for the HPIL system together with a periodontist examination, the HPIL system alone, and the consensus of two periodontists without the system. The performance of our proposed HPIL system was slightly better than the standalone periodontist examination, as displayed by the ROC graph in Figure 22, and the performance of two periodontists (sensitivity, 0.76). However, we observe that our proposed HPIL system in joint operation with a radiologist performs best out of these three scenarios. The HPIL standalone system shows slightly better performance indices (sensitivity, 0.82). On the other hand, the joint use of the HPIL system with a periodontist produced good results (sensitivity, 0.86).
A GUI was designed that will be used by the clinician for the classification of OPMDs. Figure 23 shows a graphical view of the GUI, which has been divided into a left part called “Training” and a right part called “Testing”. In the Training phase, all the dataset images are first pre-processed using the CHT and quadtree buttons. The textural patterns of these images are analyzed via the GLCM button. The results for the textural patterns of 10 different features are calculated. Once the textural patterns of all the images are calculated, the LDA button is pressed to find the best-suited features. These best-suited features are passed to the KNN classifier to validate the features selected via LDA using statistical parameters (the sensitivity, specificity, and accuracy), as shown in Table 4.
In the Testing phase, the test image is first pre-processed using CHT and quadtree. The textural pattern of this image is analyzed via the GLCM button. Once the textural patterns of all the images are calculated, the textural patterns of the features selected in training are passed to the KNN classifier to differentiate the regions, shown in the form of statistical results (sensitivity, specificity, and accuracy).
Limitations and Future Work
This paper was limited to designing a computer-aided system for the classification of OPMDs for a biopsy. The VELscope images use digital image processing based on textural analysis representations. The motivation for using GLCM was its ability to represent the textural features in the images by calculating the co-occurrence matrix. In the future, a research study will aim at the reconstruction of the GLCM co-occurrence matrix in other directions such as 45°, 90°, and 135°; the application of the algorithm on more images of specific OPMDs representing the same disorder to classify the particular OPMD; and the analysis of other textural pattern algorithms such as LBP, Fourier, and graph cut for the classification of OPMDs, and as it is challenging to capture the images of the oral cavity using the currently existing devices such as the VELscope® and Identafi®, these devices could be modified by fixing the camera inside the device, which would help the clinicians to capture the images along with the screening of patients. Furthermore, this study was limited to the detection of OPMDs and oral mucosal lesion regions with a relatively small dataset. The prognosis for OPMDs and precancerous and oral mucosal lesions with a bigger database including different ethnic groups will be a future aspect of our research.
6. Conclusions
The classification of anomalous (OML and OPMDs) and specific regions using autofluorescence images with the help of textual analysis and feature selection represents a novel way of identifying the ROI for biopsy. The method suggested in this paper involves a textural analysis of the pre-processed images to extract features, which are then ranked using LDA. Using the KNN classifier and a five-fold classifier, the best performance was obtained from the first eight features of the LDA ranking. The accuracy, sensitivity, and specificity in differentiating anomalous from standard regions of the oral cavity were 83%, 85%, and 84%, respectively. In this new area of processing oral cavity autofluorescence images, the obtained results are very encouraging. In the future, once the full dataset for OPMDs is organized, other image-processing algorithms will be used to enhance these results. Further discrimination between dysplasia and non-dysplasia in the anomalous lesions will be carried out based on GLCM texture analysis or other image-processing algorithms. These GLCM texture analyses or other image-processing algorithms for discrimination between dysplasia and non-dysplasia may further enhance the clinician’s judgment (using a COE and autofluorescence imaging) in locating the biopsy site that would best represent malignant potential.
Author Contributions
Conceptualization, approach, software, justification, formal analysis, imaging, and script—initial manuscript preparation: M.A.; examination, resources, database curation, writing—initial manuscript preparation, review, and editing: H.G., N.W., and I.F., A.K.P., M.N.S. and S.P.; data collection and subject management: A.R. and W.M.N.G.; data annotation: A.R. and R.B.Z.; clinical database and determining subject health conditions: A.R. and R.B.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This study was supported by the Ministry of Education (MOE) Malaysia High Impact Research (HIR) Grant (Grant number: UM.C/625/1/HIR/MOHE/DENT/06) and University Technology PETRONAS STRIF Funding.
Conflicts of Interest
The authors declare no conflict of interest.
Figures and Tables
Figure 1. Autofluorescence devices: (a) 3rd generation VELscope®; (b) 2nd generation VELscope®.
Figure 5. VELscope® images with Canon A620 settings; the field of view (FOV) is quite straightforward and focused, as in RGB frames.
Figure 6. VELscope® images without the Canon A620 settings; the field of view (FOV) is quite blurred and not as focused as in RGB frames.
Figure 19. Gray-level co-occurrence matrix (GLCM) texture features of all the ten parameters using 64 × 64 quadtree size.
Figure 20. Graphical analysis of each feature selected via LDA using GLCM features.
Figure 20. Graphical analysis of each feature selected via LDA using GLCM features.
Figure 22. Comparative ROC between periodontist, HPIL system, and HPIL system with the periodontist.
Clinical diagnosis and histopathological diagnosis of oral potentially malignant disorders (OPMDs) in the selected database.
Diagnosis | No. of Images | |
---|---|---|
Clinical Diagnosis | 1. Homogenous leukoplakia | 5 |
2. Non-homogenous leukoplakia | 5 | |
3. Oral lichen planus | 9 | |
4. Squamous cell carcinoma | 1 | |
5. Non-OPMDs |
10 | |
Total | 30 | |
Histopathological Diagnosis | 1. Mild dysplasia | 5 |
2. Moderate epithelial dysplasia | 4 | |
3. Oral lichen planus | 10 | |
4. Squamous cell carcinoma | 1 | |
5. Others (ulcers, abscesses, geographic tongue, epithelial hyperkeratosis, epithelial hyperplasia, median rhomboid glossitis, pyogenic granuloma) | 10 | |
Total | 30 |
Canon A620 camera settings to capture VELscope® images.
Exposure Mode | Manual |
---|---|
F-Stop | 5–4 (approx.) |
Lens aperture | 0.02 |
ISO | 1600 |
Flare | zero |
Focal point | manual |
White balance | default |
GLCM feature analysis using LDA.
Parameters | Ranked LDA Measure Values | Feature Names |
---|---|---|
1 | 4 | Variance |
2 | 3.7 | Sum average |
3 | 3 | Inverse difference moment |
4 | 2.7 | Sum variance |
5 | 2.1 | Entropy |
6 | 1.7 | Difference entropy |
7 | 1.4 | Sum entropy |
8 | 1.1 | Correlation |
9 | 0 | Contrast |
10 | 0 | Measure of co-relation |
Five-fold classification results using k-nearest neighbor (KNN).
Parameters # | Sensitivity | Specificity | Accuracy |
---|---|---|---|
1 | 40 ± 20 | 45 ± 20 | 50 ± 25 |
1, 2 | 60 ± 25 | 50 ± 20 | 50 ± 25 |
1, 2, 3 | 53 ± 20 | 55 ± 20 | 54 ± 15 |
1, 2, 3, 4 | 65 ± 18 | 70 ± 15 | 65 ± 17 |
1, 2, 3, 4, 5 | 70 ± 21 | 50 ± 24 | 60 ± 22 |
1, 2, 3, 4, 5, 6 | 72 ± 17 | 74 ± 15 | 70 ± 17 |
1, 2, 3, 4, 5, 6, 7 | 78 ± 10 | 79 ± 8 | 73 ± 13 |
1, 2, 3, 4, 5, 6, 7, 8 | 85 ± 5 | 84 ± 3 | 83 ± 5 |
1, 2, 3, 4, 5, 6, 7, 8, 9 | 50 ± 27 | 45 ± 20 | 45 ± 10 |
1, 2, 3, 4, 5, 6, 7, 8, 9, 10 | 40 ± 30 | 30 ± 15 | 39 ± 16 |
Analysis of OPMD vs. typical oral cavity region using multiple textural descriptors.
Texture Descriptors | Specificity | Sensitivity | Accuracy |
---|---|---|---|
GDP | 53 ± 20 | 55 ± 20 | 54 ± 15 |
GDP2 | 65 ± 18 | 70 ± 15 | 65 ± 17 |
GLTP | 70 ± 21 | 50 ± 24 | 60 ± 22 |
IWLD | 72 ± 17 | 74 ± 15 | 70 ± 17 |
LAP | 70 ± 21 | 50 ± 24 | 60 ± 22 |
LBP | 72 ± 17 | 74 ± 15 | 70 ± 17 |
LDIP | 53 ± 20 | 55 ± 20 | 54 ± 15 |
LDIPV | 65 ± 18 | 70 ± 15 | 65 ± 17 |
IDN | 70 ± 21 | 50 ± 24 | 60 ± 22 |
LDNP | 72 ± 17 | 74 ± 15 | 70 ± 17 |
LGIP | 65 ± 18 | 70 ± 15 | 65 ± 17 |
LGP | 70 ± 21 | 50 ± 24 | 60 ± 22 |
LPQ | 72 ± 17 | 74 ± 15 | 70 ± 17 |
LTEP | 53 ± 20 | 55 ± 20 | 54 ± 15 |
LTrP | 65 ± 18 | 70 ± 15 | 65 ± 17 |
MBC | 70 ± 21 | 50 ± 24 | 60 ± 22 |
LFD | 72 ± 17 | 74 ± 15 | 70 ± 17 |
LMP | 72 ± 17 | 74 ± 15 | 70 ± 17 |
Comparison of our purposed algorithm with existing work.
Authors | Investigation Principle | Oral Pathology | ROI Screening ⩔ | Statistical Analysis |
---|---|---|---|---|
Ganga et al. [76] | Conventional oral examination vs. VELscope® method | Evaluate the effectiveness of the VELscope® in recognizing dysplastic and/or neoplastic changes in oral mucosal lesions that were identified on conventional oral examination | Manual | Specificity = 76% |
Scheer et al. [77] | VELscope® | Oral squamous cell carcinomas | Manual | Sensitivity = 33.3% |
Farah et al. [78] | VELscope® | Oral potentially malignant disorders | Manual | Sensitivity = 30% |
Awan et al. [32] | VELscope® and conventional oral examination | Oral leukoplakia, oral erythroplakia, oral lichen planus, and oral sub-mucous fibrosis | Manual | Sensitivity = 84.1% |
Mehrotra et al. [79] | VELscope® vs. ViziLite ® | Oral squamous cell carcinomas | Manual | Sensitivity = 50% |
Our proposed algorithm | VELscope® vs. textural analysis approach | Oral squamous cell carcinomas, oral leukoplakia, oral erythroplakia, oral lichen planus, oral sub-mucous fibrosis, epithelial dysplasia lesions, and mild dysplasia lesions | Automatic | Sensitivity = 85 ± 5% |
⩔ To find region of interest (ROI): if VELscope® images are analyzed by periodontist = manual; if VELscope® image is analyzed by computer-aided system = automatic.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2020 by the authors.
Abstract
Oral mucosal lesions (OML) and oral potentially malignant disorders (OPMDs) have been identified as having the potential to transform into oral squamous cell carcinoma (OSCC). This research focuses on the human-in-the-loop-system named Healthcare Professionals in the Loop (HPIL) to support diagnosis through an advanced machine learning procedure. HPIL is a novel system approach based on the textural pattern of OML and OPMDs (anomalous regions) to differentiate them from standard regions of the oral cavity by using autofluorescence imaging. An innovative method based on pre-processing, e.g., the Deriche–Canny edge detector and circular Hough transform (CHT); a post-processing textural analysis approach using the gray-level co-occurrence matrix (GLCM); and a feature selection algorithm (linear discriminant analysis (LDA)), followed by k-nearest neighbor (KNN) to classify OPMDs and the standard region, is proposed in this paper. The accuracy, sensitivity, and specificity in differentiating between standard and anomalous regions of the oral cavity are 83%, 85%, and 84%, respectively. The performance evaluation was plotted through the receiver operating characteristics of periodontist diagnosis with the HPIL system and without the system. This method of classifying OML and OPMD areas may help the dental specialist to identify anomalous regions for performing their biopsies more efficiently to predict the histological diagnosis of epithelial dysplasia.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details




1 Center for Intelligent Medical Electronics, Department of Electronic Engineering, School of Information Science and Technology, Fudan University, Shanghai 200433, China;
2 Innovation Division Technical University of Denmark, 2800 Lyngby, Denmark;
3 Oral Diagnostic Sciences, Faculty of Dentistry, SEGi University, Jalan Teknologi, Kota Damansara, Petaling Jaya 47810, Selangor, Malaysia;
4 Oral Cancer Research and Coordinating Centre, Faculty of Dentistry, University of Malaya, Kuala Lumpur 50603, Malaysia;
5 Oral Cancer Research and Coordinating Centre, Faculty of Dentistry, University of Malaya, Kuala Lumpur 50603, Malaysia;
6 Symbiosis Centre for Applied Artificial Intelligence and CSE Dept, Symbiosis International (Deemed) University, Pune 412115, Maharashtra, India;
7 Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS, Bandar Seri Iskandar 32610, Perak, Malaysia;
8 Oral Cancer Research and Coordinating Centre, Faculty of Dentistry, University of Malaya, Kuala Lumpur 50603, Malaysia;
9 Department of Fundamental and Applied Sciences, Universiti Teknologi PETRONAS, Bandar Seri Iskandar 32610, Perak, Malaysia