This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Brain tumors are one of the most dangerous types of brain diseases that can develop due to abnormal cell growth inside the skull. Brain tumors can be categorized into two types: primary tumors and secondary tumors. Primary brain tumors account for 70% of all tumors and spread only in the brain, whereas secondary brain tumors form in other organs such as the breast, kidney, and lung before migrating to the brain. According to a study by NBTF, in US alone, around 29,000 cases of primary brain tumor are diagnosed each year, resulting in the death of 13,000 people [1]. Similarly, in the United Kingdom, about 42,000 people with primary brain tumors die each year. Glioma, Meningioma, and Pituitary tumors are the most prevalent brain tumors. Glioma tumor is caused by unusual growth in Glial cells that constitute 80% of the brain. Among all primary tumors, it has the highest fatality rate. Meningioma tumors develop in the brain’s protective membrane, the meninges spinal cord. In contrast, the pituitary tumor develops in the pituitary gland. This gland produces various necessary hormones. Although the pituitary tumor is benign, it can cause hormonal deficiencies and irreparable damage to vision [2]. Hence, an early and accurate diagnosis of brain tumors is necessary to protect patients from damaging effects [3]. Depending on their objective, brain tumors can be diagnosed using various medical imaging technologies. Ultrasonography (US), magnetic resonance imaging (MRI), and computed tomography (CT) are three of the widely used techniques [4]. The most prevalent noninvasive imaging technology is magnetic resonance imaging (MRI), which does not emit any harmful ionizing radiation during the examination like X-rays. Furthermore, it generates clear images of soft tissues and can acquire modalities like FLAIR, T1, and T2, using a variety of parameters [5].
Proper identification of tumor type is a difficult task as the tumors usually vary in shape, intensity, size, and location. Usually, the medical professionals visually inspect the images and meticulously mark out the tumor regions in the images. Because of the surrounding healthy tissues, tumor borders are frequently blurred. As a result, the manual identification process via optical inspection is time-consuming and can cause misinterpretation of the tumor. Furthermore, manual tumor detection relies heavily on the radiologist’s experience [6]. It should also be mentioned that the human eye cannot distinguish between distinct shades of grey shown in MRI scans. Other prominent reasons for tumor misinterpretation include fatigued radiologists or noisy MRIs caused by variations in imaging devices. Thus, automated systems are appropriate when radiologists want to visually evaluate the depth of the tumor or identify the type of tumor to reduce the likelihood of biopsy [7, 8]. Various researchers have proposed CAD-based brain tumor detection methods. However, the limitation of traditional ML-based algorithms is that they use a hand-crafted feature extraction strategy. The features are extracted from training images before classification [9].
Brain tumor classification techniques can be divided into machine learning- (ML-) based methods and deep learning- (DL-) based methods. The ML-based systems employ handcrafted feature extraction and manual segmentation before classification that is time-consuming and error-prone. These methods typically require the assistance of an expert with extensive experience to discover optimal feature extraction and segmentation algorithms for proper tumor identification. As a result, when working with larger databases, the performance of these systems is prone to errors [10]. Meanwhile, DL-based algorithms perform these steps automatically and have proven incredibly useful in various applications, including medical image analysis. One of the most prominent DL models, i.e., convolutional neural network (CNN), is widely employed because of its robust performance and weight-sharing nature. It can extract low- level and high-level features from training data automatically. Hence, these techniques have drawn the interest of scientists and researchers [11, 12].
This paper is a continuation of research presented in Kibriya et al. [9] that compares the performance of two transfer learning architectures such as ResNet18 and GoogLeNet to classify brain tumors from MRI images. The deep features were classified using end-to-end CNN models as well as ECOC based SVM. This paper proposes an automated method based on the fusion of deep features obtained via three well-known CNNs, AlexNet, ResNet18, and GoogLeNet, for brain tumor identification and classification. Feature fusion is a method that combines multiple low and high-level features into a single feature vector, thus increasing the discrimination performance of model by eliminating the need of utilizing feature vector from a single model. The motivation behind using a feature fusion-based technique is to produce informative and discriminative features from MRIs, which are critical for accurate tumor classification. To demonstrate the efficacy of the suggested method, we used multiple quantitative measures to assess our model on a well-known brain tumor dataset [13]. The primary contributions of the proposed system are as follows:
(i) We designed a completely automated hybrid system that uses both (a) transfer learned CNN models for deep feature extraction and (b) ML classifiers for effective classification of the type of brain tumors.
(ii) The proposed method consists of five core steps such as (a) image intensity normalization, (b) extensive dataset augmentation, (c) feature extraction via multiple CNNs, that is, AlexNet, ResNet18, and GoogLeNet, (d) fusion of deep features vector that achieves cutting-edge performance for classifying brain tumors from MR images, and (e) tumor type classification via SVM and KNN.
(iii) The concatenation of deep features in a single vector enhances the discriminating power. It enables effective tumor classification under various challenges such as complex background, fuzzy tumor boundaries, varying tumor location, and other problems in MRI artifacts.
(iv) The obtained results show the robustness of the proposed approach compared to existing systems.
The paper is arranged as follows. Section 2 analyzes related studies. Section 3 describes the proposed method. Section 4 and Section 5 are dedicated to experiments and results, respectively. The proposed work is concluded in Section 6.
2. Literature Review
Early tumor identification and classification of brain tumors are necessary for the effective treatment of a patient. Experts can now provide better treatment to patients via automated healthcare systems because of remarkable technology improvements. The researchers have proposed studies employing ML and DL based algorithms to solve medical image diagnosis problems [14]. One of the most famous DL models is CNN, which has achieved groundbreaking results in different fields, including image processing. CNN-based systems can efficiently diagnose brain tumors and assist healthcare providers in determining treatment decisions for patients. According to the statistics of 2016, more than 200 DL-based researches on medical images were proposed, and 190 of those employed the CNNs [11]. Some of the very popular CNNs such as AlexNet [15], VGG [16]
Mzoughi et al. [18] demonstrated 3D CNN architecture for glioma brain tumor classification into low-grade gliomas (LGG) and high-grade gliomas (HGG) utilizing the entire volumetric T1-Gado MRI sequence from Brats 2018 dataset. Using small kernels, the architecture combined local and global contextual information with lower weights, based on a 3D convolutional layer and a deep network. The system achieved 96.49% accuracy. Maqsood et al. [19] suggested a brain tumor detection method employing edge detection and U-NET model. The tumor segmentation framework enhances image contrast and performs edge detection fuzzy logic. The features are extracted from decaying subband images and then classified using the U-NET architecture, which detects the presence of meningioma in brain images. Togacar et al. [20] developed BrainMRNet using attention modules and the hypercolumn method. Initially, the images were preprocessed before being sent to the attention modules. Attention modules determine the key regions of an image and send the image to convolutional layers. Hypercolumn is one of the important techniques used by the BrainMRNet model in the convolutional layers. The features extracted from each layer are maintained by the array structure in the final layer using this technique. The system achieved an accuracy of 96.05%. The authors in Khawaldeh et al. [21] suggested a CNN model to detect brain tumors and Glioma tumors by improving pretrained architecture and achieving 91% overall accuracy. Despite the tremendous amount of work in this field, developing a good and practical technique for classifying brain MR images still requires additional research. The main limitation of the research in [18–21] is that they only perform binary classification of brain tumors and disregard multiclass classification, requiring additional analysis to determine the kind of tumor.
Several recent researches employing transfer learning-based methods to detect brain tumors have been proposed. For example, the authors in Sajjad et al. [22] demonstrated a multimodal tumor classification system based on CNN. They initially segmented the MR images using Input Cascade CNN and classified them using a fine-tuned VGG-19 with 94.5% accuracy. However, the framework is computationally inefficient as it performs segmentation and classification via CNNs. Ari et al. [23] fused deep features obtained from AlexNet and VGG16. The fused feature vector was then classified via Extreme Learning Machine (ELM). The study was conducted on MRI images from publicly available Figshare, Rider, and REMBRANDT datasets. The system achieved 96.6% accuracy. Noreen et al. [24] extracted deep features using VGG16, VGG19, and AlexNet and classified the features via ensemble classifiers. The system achieved the highest accuracy of 94.3%. Swati et al. [25] classified brain tumor MRI images using fine-tuned AlexNet and VGG with an accuracy of 94.8%. The authors in Saxena et al. [26] employed ResNet, Inception-V3, and VGG-16 and achieved the highest accuracy of 95% via ResNet. However, the techniques obtained a low overall performance and need to be tested before real-time deployment.
The authors in Abiwinanda et al. [27], on the other hand, suggested five CNN designs to classify brain tumors and obtained the highest validation accuracy of 84.1%. It may be noted that very simplistic CNNs cannot extract deep high-level features, resulting in poor overall accuracy. Alanazi et al. [28] presented a brain tumor classification framework employing novel 22 layered CNN architectures and achieved the highest accuracy of 96.8%. However, the proposed technique is trained and evaluated on limited imaging samples. Khan et al. [29] proposed a Hierarchical Deep Learning-Based Brain Tumor classification method. The study was conducted on MRI images from Kaggle dataset and achieved 92.13% classification accuracy. However, the system needs to be tested before deploying for clinical setup for brain tumor classification due to low overall accuracy. Anaraki et al. [30] employed the Genetic Algorithms (GA) to find an optimal CNN architecture with lesser computation cost for the classification of brain tumors. They obtained 94.2% accuracy to classify Glioma, Meningioma, and Pituitary tumors from MRI images. However, GA could not find optimal CNN architecture, thus resulting in poor overall accuracy. The authors in Raja [31] employed Bayesian fuzzy clustering (BFC) technique for image segmentation, nonlocal mean filter for image denoising and scattering transform, information-theoretic measurements, and wavelet packet Tsallis entropy for feature extraction and a hybrid DAE strategy for brain tumor classification. However, this technique takes a long time to compute and is computationally inefficient. Afshar et al. [32] used Capsule Networks to identify and classify brain lesions with 90.89%. It should be noted that CapsNets are particularly sensitive to image backgrounds and perform better when segmented images are used to train the model. As a result, the architecture is complicated.
3. Proposed Methodology
This paper proposes a novel MRI-based brain tumor identification and classification method employing a fusion of deep CNN features. The workflow diagram is presented in Figure 1. The MRI images are normalized and augmented before being fed into three separate CNN models for feature extraction. The obtained deep feature vectors are fused in a single feature vector and classified via SVM and KNN. The proposed method is robust and efficient and can be utilized to accurately classify different types of brain tumors such as Pituitary, Glioma, and Meningioma.
[figure omitted; refer to PDF]
In this study, we employed CNN architectures, namely, AlexNet [15], GoogLeNet [17]
[figures omitted; refer to PDF]
Table 2
Parameters of deep learning models.
CNN | Input size | Depth | Parameters (M) |
GoogLeNet | 224 × 224 | 22 | 7 |
ResNet18 | 224 × 224 | 18 | 11 |
AlexNet | 227 × 227 | 8 | 61 |
3.4. Fusion of Deep CNN Features
The success of ML classifier is heavily reliant on an input feature vector; thus, developing an algorithm to generate informative and discriminative features from MRIs is crucial for accurate tumor classification. In this phase, we fuse the deep features obtained from the transfer learned CNNs in the previous phase. Feature fusion is a technique that merges numerous features from several different models into a single feature vector, thus avoiding the need to utilize a single feature vector obtained from a model with low overall performance. This integration of features will likely improve classification results as the new feature vector contains more information about the MR images than a single vector [8]. The deep CNN features obtained from two homogenous architectures have identical feature spaces. As a result, the fusion of these features may contain repetitive feature space, thus lacking diversity. Hence, we used heterogeneous CNN architectures with varying architectural design and depth to extract diverse low level and high level features from MR images. The novel fused feature vector contains more information than the independent feature vectors. Figure 7 depicts the feature fusion procedure employed in the proposed study. Each independent feature vector has three feature spaces according to the number of classes in the dataset; hence, the fused feature vector consists of 9 features.
[figure omitted; refer to PDF]
Figure 9 illustrates the exploratory data analysis to understand the dataset class distribution. The pie chart shows that Glioma samples compose 46% of the entire dataset. Meningioma and Pituitary images account for only 23% and 30% of the overall dataset, respectively. The chart indicates class imbalance issues; hence, in such conditions, Recall, F1-Score, and Precision are more appropriate metrics compared to accuracy.
[figure omitted; refer to PDF]4.2. Experimental Settings
This study employed transfer learning CNNs such as ResNet18, AlexNet, and GoogLeNet as deep feature extractors. For classification, the deep features were fused and supplied to SVM and KNN. All the experiments are performed on an Intel Core i5 processor with an 8 GB RAM using Matlab R2021a. The hyperparameters of the CNN and the values are shown in Table 4.
Table 4
Training parameters for CNNs.
Hyper-parameters | Values |
Epoch | 30 |
Mini-batch size | 10 |
Optimizer | SGDM |
Learning rate | 1e − 4 |
We trained the CNNs using 30 epochs. This is an epoch when an entire dataset is passed forward and backward through the architecture only once. The epoch is too large to process all at once, so it is usually broken into smaller batches. The total number of training examples in a single batch (or mini-batch) is referred to as a batch size/mini-batch size (such as 10 in our case) [41]. The CNNs in the proposed frameworks are trained on a 1e − 4 learning rate. The learning rate determines how quickly an algorithm learns the values of a parameter estimate. It may be noted that higher learning rates frequently result in a suboptimal set of weights [42]. Moreover, we trained the CNNs using stochastic gradient descent momentum (SGDM) optimizer. The SGD optimizes the network parameters to achieve optimal loss function by taking incremental steps towards negative gradient at each iteration. SGD can oscillate along the steepest descent path to the best solution. One method for reducing this oscillation is to include a momentum term in the parameter update [43], as shown in the following equation:
4.3. Evaluation Metrics
When designing automated systems, evaluating the model’s performance is vital because the main purpose of such a model is to forecast unanticipated data reliably; hence, analyzing the training and validation test set reveals the generalization ability of the framework. Usually, a classification model is assessed using a confusion matrix, a simple cross-tabulation of actual and expected observations for each class. Various classification measures based on the confusion matrix, such as accuracy, precision, recall, and f1-score, are used as a benchmark to assess the model’s performance. The classification accuracy is a popular metric to summarize the model’s overall performance. The f1-score, on the other hand, combines precision and recall into a single statistic that contains both features. The recall is calculated in equation (3), Precision in equation (4), Accuracy in equation (5)
4.4. Proposed Method Results
This section discusses the proposed method results. Table 5 compares the accuracy values obtained via independent CNN feature vectors and novel fused feature vectors. The feature vectors from independent CNN architectures such as ResNet18 and AlexNet achieved the maximum accuracy of 98%, while GoogLeNet produced an accuracy of 97.6%. On the other hand, the fused feature vector obtained an accuracy of 99.7% and outperformed the independent feature vectors.
Table 5
Accuracy % obtained from independent and fused feature vectors.
Feature vector/Classifiers | Accuracy % | |
SVM | KNN | |
GoogLeNet | 97.6 | 97.6 |
ResNet18 | 98.0 | 97.7 |
AlexNet | 97.9 | 98.0 |
Fused | 99.7 | 99.7 |
The performance of our proposed technique is also evaluated using other metrics such as F1-Score, recall, and precision, as shown in Figure 10. The classifiers achieved precision, recall, and F1-Score of 0.99, 0.1, and 0.99, respectively. These metrics indicate our proposed method’s robustness despite the dataset’s class imbalance due to a fused feature vector with more robust and discriminative features than an independent feature vector. Table 6 shows metrics such as Recall, Precision, and F1-Score obtained from the confusion matrix class-wise.
[figure omitted; refer to PDF]Table 6
Class-wise evaluation metrics obtained from confusion matrix.
Tumor type | F1-score | Recall | Precision |
Glioma | 1.0 | 1.0 | 1.0 |
Meningioma | 0.99 | 1.0 | 0.99 |
Pituitary | 1.0 | 1.0 | 1.0 |
Average | 0.99 | 1.0 | 0.99 |
Figure 11(a) shows the confusion matrix obtained from SVM. The X-axis of the matrix represents the predicted class, while the Y-axis represents the true class. The matrix indicates that SVM correctly classified 2132 Glioma samples, 1051 Meningioma samples, and 1393 Pituitary samples. The confusion matrix obtained on the fused feature vector via KNN is shown in Figure 11(b). The matrix illustrates that KNN correctly identified 2130 Glioma samples, 1052 Meningioma samples, and 1393 Pituitary samples. Both classifiers obtained an accuracy of 99.7%.
[figures omitted; refer to PDF]
Figure 12 shows Receiver Operating Curves (ROC) for all three classes: Glioma, Meningioma, and Pituitary. ROC analysis was originally developed during World War II to analyze noise in the radar signals [44]. The use of ROC curves for assessing the performance of medical diagnostic systems has grown in popularity over the last few decades. The curve indicates the trade-off between specificity and sensitivity. It is more efficient when the curve is closer to the upper left corner. One of the best properties of ROC is that the accuracy indices generated from the analysis are not influenced by fluctuations caused by arbitrarily chosen decision criteria or cut-offs [45]. The area under the ROC curve determines the discriminative capacity showing how efficiently it performs in a certain scenario [46]. An excellent model has an AUC close to one, indicating a high level of separability between the classes.
[figures omitted; refer to PDF]
5. Result Comparison with State-of-the-Art Techniques
Recent advancements in medical image analysis tools have provided health practitioners more convenience with detecting diseases at an early stage. Such advancements are assisting them in various fields of medicine, including disease identification, therapy, and quick decision-making for clinical applications. Every day, hospitals generate a large amount of medical data. Medical informatics research assists doctors and scientists in their search for the best solutions for making good use of these ever-increasing volumes of data [47, 48].
Early detection and appropriate treatment options are essential to treat brain tumor illnesses effectively. Those treatment options are determined by the stage of tumor, type, and grade at the diagnosis time. The conventional identification systems employs basic ML-oriented algorithms that extract limited features [7, 49]. This paper presents a novel feature fusion-based strategy to classify brain tumors from MR images accurately.
Table 7 gives a comprehensive comparison of our proposed tumor detection and classification algorithms with the existing methodologies using accuracy as a metric. Afshar et al. [32] classified brain tumors with 90.8% using a CapsNet architecture. However, CapsNet architectures are extremely sensitive to image background, severely impacting the performance. Furthermore, they gave tumor boundaries extra input for improved results, necessitating manual tumor localization prior to classification. Anaraki et al. [30] employed GA to determine optimal CNN design and achieved an accuracy of 94.2%. However, GA was unable to find best CNN for brain tumor diagnosis, thus resulting in low overall accuracy. Kang et al. [8] fused deep features from DenseNet169, Inception-v3, and ResNeXt50 to classify brain tumors. The system obtained a high accuracy of 98.5%. Ari et al. [23] fused deep features obtained from AlexNet and VGG16, then classified via ELM, and achieved 96.6% accuracy, whereas our proposed brain tumor classification approach attained the best accuracy of 99.7% and performed better than the existing methodologies.
Table 7
Comparison of proposed techniques with the existing systems.
Reference | Technique | Accuracy% |
Anarkari et al. [30] | CNN | 94.2 |
Afshar et al. [32] | CapsNet | 90.8 |
Sejuti and Islam [50] | CNN + SVM | 97.1 |
Kang et al. [8] | DenseNet169 + Inception-v3 + ResNeXt50 | 98.5 |
Ari et al. [23] | AlexNet + VGG16 | 96.6 |
Proposed | AlexNet + GoogLeNet + ResNet18 | 99.7 |
Table 8 compares the performance of the proposed feature fusion-based method with non-DL-based techniques. The authors in Thejaswini et al. [52] used the Adaptive Regularized Kernel-based Fuzzy C-Means Clustering (ARKFCM) approach to segment images and extract statistical features from the segmented area. Using the SVM, the system achieved a success rate of 91.4%. Kaplan et al. [53] suggested an approach that uses Local Binary Pattern (LBP) as a feature extraction technique and KNN as a classification technique. The system achieved the highest accuracy of 95.5%. Chen et al. [28] evaluated the effectiveness of various feature extraction methods such as density histogram, grey level cooccurrence matrix (GLCM), and Bag of Word (BoW). They attained the highest accuracy of 91.4% with SVM and KNN as classifiers and BoW as a feature extraction method. The traditional approaches usually rely on manual tumor segmentation and feature extraction, which is both time-consuming and error-prone. To find an optimal solution, these techniques need to be tried and tested by a knowledgeable expert.
Table 8
Comparison with traditional ML techniques.
Reference | Technique/s | Accuracy (%) |
Chen et al. [51] | BoW, GLCM, SVM, and KNN | 91.4 |
Thejaswini et al. [52] | ARKFCM and SVM | 91.4 |
Kaplan et al. [53] | LBP and KNN | 95.5 |
Proposed | CNN, SVM, and KNN | 99.7 |
CNNs have made significant advances in image processing, allowing any classification or segmentation operation to be performed precisely, crucial in biomedical applications. This research proposes a novel brain tumor classification technique based on the fusion of deep features extracted using three distinct CNN architectures. The proposed brain tumor classification framework is an efficient and segmentation-free approach that employs a hybrid feature set. The fusion of multiple feature vectors produces a more discriminative feature representation than an independent vector. Usually, the features extracted from different CNN architectures are more robust and descriptive than those obtained from homogenous CNN architectures or a single CNN architecture because of the diverse range of low-level and high-level features.
The dataset used in this study, however, is imbalanced. In general, a well-balanced data set with an equal class distribution yields better prediction accuracy. It is worth noting that a well-balanced dataset makes learning easier for the classification system. The dataset used in this study is available publicly; thus, the class imbalance is obvious. Hence, the dataset used in this study is extensively augmented to deal with the class imbalance problems. Moreover, due to the imbalanced dataset, the proposed framework is also assessed using recall, precision, and f1-score. The results prove that the proposed framework can be deployed in a clinical setup for real-time diagnosis of brain tumors.
6. Conclusion and Future Work
This paper presents a novel deep feature fusion-based framework to classify brain tumors from MR images. The proposed framework extracts a wide range of low-level and high-level features from AlexNet, GoogLeNet, and ResNet18 architectures with varying depth and design. These features are then merged using a serial fusion approach to generate a single vector. The novel feature vector contains robust combination features classified using SVM and KNN. The proposed method is trained and evaluated on 15,320MR images and obtained accuracy of 99.7%, recall value of 1.0, a precision score of 0.99, and 0.99 f1-score. The novel feature vector outperformed the independent CNN feature vectors. Moreover, the feature fusion strategy aids in overcoming the drawbacks of a single CNN model, thus resulting in higher performance, particularly for larger datasets. The obtained results prove the efficacy and robustness of the proposed method in brain tumor classification. Hence, the proposed framework can assist radiologists in detecting and classifying brain tumors accurately. In the future, we will explore other CNNs such as VGG, DenseNet, and machine learning classifiers such as Random Forests and Ensemble learning. Moreover, additional MRI datasets will be gathered with other tumor categories and different imaging modalities for classification.
Acknowledgments
This study was supported by Taif University Researchers Supporting Project number (TURSP-2020/215), Taif University, Taif, Saudi Arabia.
[1] M. Akil, R. Saouli, R. Kachouri, "Fully automatic brain tumor segmentation with deep learning-based selective attention using overlapping patches and multi-class weighted cross-entropy," Medical Image Analysis, vol. 63, 2020.
[2] J. Komninos, V. Vlassopoulou, D. Protopapa, S. Korfias, G. Kontogeorgos, D. E. Sakas, N. C. Thalassinos, "Tumors metastatic to the pituitary gland: case report and literature review," Journal of Clinical Endocrinology & Metabolism, vol. 89 no. 2, pp. 574-580, DOI: 10.1210/jc.2003-030395, 2004.
[3] H. Habib, R. Amin, B. Ahmed, A. Hannan, "Hybrid algorithms for brain tumor segmentation, classification and feature extraction," Journal of Ambient Intelligence and Humanized Computing, vol. 119,DOI: 10.1007/s12652-021-03544-8, 2021.
[4] M. N. Ullah, Y. Park, G. B. Kim, C. Kim, C. Park, H. Choi, J.-Y. Yeom, "Simultaneous acquisition of ultrasound and gamma signals with a single-channel readout," Sensors, vol. 21 no. 4,DOI: 10.3390/s21041048, 2021.
[5] S. Bauer, R. Wiest, L.-P. Nolte, M. Reyes, "A survey of MRI-based medical image analysis for brain tumor studies," Physics in Medicine and Biology, vol. 58 no. 13, pp. R97-R129, DOI: 10.1088/0031-9155/58/13/r97, 2013.
[6] F. Jamil, M. A. Iqbal, R. Amin, D. Kim, "Adaptive thermal-aware routing protocol for wireless body area network," Electronics, vol. 8 no. 1,DOI: 10.3390/electronics8010047, 2019.
[7] S. Deepak, P. Ameer, "Brain tumor classification using deep CNN features via transfer learning," Computers in Biology and Medicine, vol. 111, 2020.
[8] J. Kang, Z. Ullah, J. Gwak, "MRI-based brain tumor classification using ensemble of deep features and machine learning classifiers," Sensors, vol. 21 no. 6,DOI: 10.3390/s21062222, March 2021.
[9] H. Kibriya, M. Masood, M. Nawaz, R. Rafique, S. Rehman, "Multiclass brain tumor classification using convolutional neural network and support vector machine," ,DOI: 10.1109/majicc53071.2021.9526262, .
[10] S. Khalid, T. Khalil, S. Nasreen, "A survey of feature selection and feature extraction techniques in machine learning," pp. 372-378, DOI: 10.1109/sai.2014.6918213, .
[11] G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, J. A. W. M. van der Laak, B. van Ginneken, C. I. Sánchez, "A survey on deep learning in medical image analysis," Medical Image Analysis, vol. 42, pp. 60-88, DOI: 10.1016/j.media.2017.07.005, 2017.
[12] R. Amin, M. A. Al Ghamdi, S. H. Almotiri, M. Alruily, "Healthcare techniques through deep learning: issues, challenges and opportunities," IEEE Access, vol. 9, 2021.
[13] C. Jun, "Brain Tumor Dataset," 2017. https://figshare.com/articles/brain_tumor_dataset/1512427
[14] A. Gumaei, M. M. Hassan, M. R. Hassan, A. Alelaiwi, G. Fortino, "A hybrid feature extraction method with regularized extreme learning machine for brain tumor classification," IEEE Access, vol. 7,DOI: 10.1109/access.2019.2904145, 2019.
[15] A. Krizhevsky, I. Sutskever, G. E. Hinton, "Imagenet classification with deep convolutional neural networks," Advances in Neural Information Processing Systems, vol. 25, pp. 1097-1105, 2012.
[16] K. Simonyan, A. Zisserman, "Very deep convolutional networks for large-scale image recognition," 2014. https://arxiv.org/abs/1409.1556
[17] C. Szegedy, W. Liu, "Going deeper with convolutions," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR),DOI: 10.1109/cvpr.2015.7298594, .
[18] H. Mzoughi, I. Njeh, A. Wali, M. B. Slima, A. BenHamida, C. Mhiri, K. B. Mahfoudhe, "Deep multi-scale 3D convolutional neural network (CNN) for MRI gliomas brain tumor classification," Journal of Digital Imaging, vol. 33 no. 4, pp. 903-915, DOI: 10.1007/s10278-020-00347-9, 2020.
[19] S. Maqsood, R. Damasevicius, F. M. Shah, "An efficient approach for the detection of brain tumor using fuzzy logic and U-NET CNN classification," Computational Science and Its Applications - ICCSA 2021, pp. 105-118, DOI: 10.1007/978-3-030-86976-2_8, 2021.
[20] M Toğaçar, B Ergen, Z Cömert, "BrainMRNet: brain tumor detection using magnetic resonance images with a novel convolutional neural network model," Medical Hypotheses, vol. 134,DOI: 10.1016/j.mehy.2019.109531, 2020.
[21] S. Khawaldeh, U. Pervaiz, A. Rafiq, R. S. Alkhawaldeh, "Noninvasive grading of glioma tumor using magnetic resonance imaging with convolutional neural networks," Applied Sciences, vol. 8 no. 1, 2018.
[22] M. Sajjad, S. Khan, K. Muhammad, W. Wu, A. Ullah, S. W. Baik, "Multi-grade brain tumor classification using deep CNN with extensive data augmentation," Journal of computational science, vol. 30, pp. 174-182, DOI: 10.1016/j.jocs.2018.12.003, 2019.
[23] A. Arı, O. F. Alcin, D. Hanbay, "Brain MR image classification based on deep features by using extreme learning machines," Biomedical Journal of Scientific and Technical Research, vol. 25 no. 3, 2020.
[24] N. Noreen, S. Palaniappan, A. Qayyum, I. Ahmad, M. O. Alassafi, "Brain tumor classification based on fine-tuned models and the ensemble method," Computers, Materials & Continua, vol. 67 no. 3, pp. 3967-3982, DOI: 10.32604/cmc.2021.014158, 2021.
[25] Z. N. K. Swati, Q. Zhao, M. Kabir, F. Ali, Z. Ali, S. Ahmed, J. Lu, "Brain tumor classification for MR images using transfer learning and fine-tuning," Computerized Medical Imaging and Graphics, vol. 75, pp. 34-46, DOI: 10.1016/j.compmedimag.2019.05.001, 2019.
[26] P. Saxena, A. Maheshwari, S. Maheshwari, "Predictive modeling of brain tumor: a Deep learning approach," Innovations in Computational Intelligence and Computer Vision, pp. 275-285, DOI: 10.1007/978-981-15-6067-5_30, 2021.
[27] N. Abiwinanda, M. Hanif, S. T. Hesaputra, A. Handayani, T. R. Mengko, "Brain tumor classification using convolutional neural network," pp. 183-189, DOI: 10.1007/978-981-10-9035-6_33, .
[28] M. F. Alanazi, M. U. Ali, S. J. Hussain, A. Zafar, M. Mohatram, M. Irfan, R. AlRuwaili, M. Alruwaili, N. H. Ali, A. M. Albarrak, "Brain tumor/mass classification framework using magnetic-resonance-imaging-based isolated and developed transfer deep-learning model," Sensors, vol. 22 no. 1,DOI: 10.3390/s22010372, 2022.
[29] A. H. Khan, "Intelligent model for brain tumor identification using deep learning," Applied Computational Intelligence and Soft Computing, vol. 2022, 2022.
[30] A. K. Anaraki, M. Ayati, F. Kazemi, "Magnetic resonance imaging-based brain tumor grades classification and grading via convolutional neural networks and genetic algorithms," Biocybernetics Biomedical Engineering, vol. 39 no. 1, pp. 63-74, 2019.
[31] P. M. Siva Raja, A. V. rani, "Brain tumor classification using a hybrid deep autoencoder with Bayesian fuzzy clustering-based segmentation approach," Biocybernetics and Biomedical Engineering, vol. 40 no. 1, pp. 440-453, DOI: 10.1016/j.bbe.2020.01.006, 2020.
[32] P. Afshar, K. N. Plataniotis, A. Mohammadi, "Capsule networks for brain tumor classification based on mri images and coarse tumor boundaries," Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1368-1372, DOI: 10.1109/icassp.2019.8683759, .
[33] N. Ghassemi, A. Shoeibi, M. Rouhani, "Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images," Biomedical Signal Processing and Control, vol. 57,DOI: 10.1016/j.bspc.2019.101678, 2020.
[34] S. Akçay, M. E. Kundegorski, M. Devereux, T. P. Breckon, "Transfer learning using convolutional neural networks for object classification within x-ray baggage security imagery," pp. 1057-1061, .
[35] K. He, X. Zhang, S. Ren, J. Sun, "Deep residual learning for image recognition," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, DOI: 10.1109/cvpr.2016.90, .
[36] H. Kibriya, R. Rafique, W. Ahmad, S. Adnan, "Tomato leaf disease detection using convolution neural network," Proceedings of the IEEE International Bhurban Conference on Applied Sciences and Technologies (IBCAST), pp. 346-351, DOI: 10.1109/ibcast51254.2021.9393311, .
[37] S. Suthaharan, "Support vector machine," Machine Learning Models and Algorithms for Big Data Classification, pp. 207-235, DOI: 10.1007/978-1-4899-7641-3_9, 2016.
[38] C. Cortes, V. Vapnik, "Support-vector networks," Machine Learning, vol. 20 no. 3, pp. 273-297, DOI: 10.1007/bf00994018, 1995.
[39] Y. Ma, G. Guo, Support Vector Machines Applications, 2014.
[40] L. Peterson, "K-nearest neighbor," Scholarpedia, vol. 4 no. 2,DOI: 10.4249/scholarpedia.1883, 2009.
[41] J. Brownlee, "What is the difference between a batch and an epoch in a neural network," Machine Learning Mastery, vol. 20, 2018.
[42] J. Brownlee, "How to grid search hyperparameters for deep learning models in python with keras," 2016. https://machinelearningmastery.com/grid-search-hyperparameters-deep-learning-models-python-keras/
[43] K. P. Murphy, Machine Learning: A Probabilistic Perspective, 2012.
[44] K. H. Zou, A. J. O’Malley, L. Mauri, "Receiver-operating characteristic analysis for evaluating diagnostic tests and predictive models," Circulation, vol. 115 no. 5, pp. 654-657, DOI: 10.1161/circulationaha.105.594929, 2007.
[45] K. Hajian-Tilaki, "Receiver operating characteristic (ROC) curve analysis for medical diagnostic test evaluation," Caspian journal of internal medicine, vol. 4 no. 2, 2013.
[46] S. Ekelund, "ROC curves-what are they and how are they used?," Point of Care: The Journal of Near-Patient Testing & Technology, vol. 11 no. 1, pp. 16-21, DOI: 10.1097/poc.0b013e318246a642, 2012.
[47] F. Jiang, Y. Jiang, H. Zhi, Y. Dong, H. Li, S. Ma, Y. Wang, Q. Dong, H. Shen, Y. Wang, "Artificial intelligence in healthcare: past, present and future," Stroke and Vascular Neurology, vol. 2 no. 4, pp. 230-243, DOI: 10.1136/svn-2017-000101, 2017.
[48] H. H. Sultan, N. M. Salem, W. Al-Atabany, "Multi-classification of brain tumor images using deep neural network," IEEE Access, vol. 7,DOI: 10.1109/access.2019.2919122, 2019.
[49] M. A. Khan, I. Ashraf, M. Alhaisoni, R. Damaševičius, R. Scherer, A. Rehman, S. A. C. Bukhari, "Multimodal brain tumor classification using deep learning and robust feature selection: a machine learning application for radiologists," Diagnostics, vol. 10 no. 8,DOI: 10.3390/diagnostics10080565, 2020.
[50] Z. A. Sejuti, M. S. Islam, "An efficient method to classify brain tumor using CNN and SVM," pp. 644-648, DOI: 10.1109/icrest51555.2021.9331060, .
[51] J. Cheng, W. Huang, S. Cao, R. Yang, W. Yang, Z. Yun, Z. Wang, Q. Feng, "Correction: enhanced performance of brain tumor classification via tumor region augmentation and partition," PloS One, vol. 10 no. 12,DOI: 10.1371/journal.pone.0144479, 2015.
[52] P. Thejaswini, M. B. Bhat, M. K. Prakash, "Detection and classification of tumour in brain MRI," Int. J. Eng. Manufact.(IJEM), vol. 9 no. 1, pp. 11-20, 2019.
[53] K. Kaplan, Y. Kaya, M. Kuncan, H. M. Ertunç, "Brain tumor classification using modified local binary patterns (LBP) feature extraction methods," Medical Hypotheses, vol. 139,DOI: 10.1016/j.mehy.2020.109696, 2020.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2022 Hareem Kibriya et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
Brain tumors are difficult to treat and cause substantial fatalities worldwide. Medical professionals visually analyze the images and mark out the tumor regions to identify brain tumors, which is time-consuming and prone to error. Researchers have proposed automated methods in recent years to detect brain tumors early. These approaches, however, encounter difficulties due to their low accuracy and large false-positive values. An efficient tumor identification and classification approach is required to extract robust features and perform accurate disease classification. This paper proposes a novel multiclass brain tumor classification method based on deep feature fusion. The MR images are preprocessed using min-max normalization, and then extensive data augmentation is applied to MR images to overcome the lack of data problem. The deep CNN features obtained from transfer learned architectures such as AlexNet, GoogLeNet, and ResNet18 are fused to build a single feature vector and then loaded into Support Vector Machine (SVM) and K-nearest neighbor (KNN) to predict the final output. The novel feature vector contains more information than the independent vectors, boosting the proposed method’s classification performance. The proposed framework is trained and evaluated on 15,320 Magnetic Resonance Images (MRIs). The study shows that the fused feature vector performs better than the individual vectors. Moreover, the proposed technique performed better than the existing systems and achieved accuracy of 99.7%; hence, it can be used in clinical setup to classify brain tumors from MRIs.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Department of Computer Science, University of Engineering and Technology, Taxila, Pakistan
2 Durma College of Science and Humanities Shaqra University, Shaqra 11961, Saudi Arabia
3 Department of Information Technology, College of Computer and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
4 Department of Information Technology, Al Baha University, Al Bahah, Saudi Arabia