Skip to main content

ORIGINAL RESEARCH article

Front. Comput. Neurosci., 02 September 2022
Volume 16 - 2022 | https://doi.org/10.3389/fncom.2022.1005617

Ensemble deep learning for brain tumor detection

  • 1College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, AlKharj, Saudi Arabia
  • 2Department of Accounting and Information Systems, College of Business and Economics, Qatar University, Doha, Qatar
  • 3Department of Computer Science, COMSATS University, Islamabad, Pakistan
  • 4Department of Computer Science and Software Engineering, International Islamic University, Islamabad, Pakistan

With the quick evolution of medical technology, the era of big data in medicine is quickly approaching. The analysis and mining of these data significantly influence the prediction, monitoring, diagnosis, and treatment of tumor disorders. Since it has a wide range of traits, a low survival rate, and an aggressive nature, brain tumor is regarded as the deadliest and most devastating disease. Misdiagnosed brain tumors lead to inadequate medical treatment, reducing the patient's life chances. Brain tumor detection is highly challenging due to the capacity to distinguish between aberrant and normal tissues. Effective therapy and long-term survival are made possible for the patient by a correct diagnosis. Despite extensive research, there are still certain limitations in detecting brain tumors because of the unusual distribution pattern of the lesions. Finding a region with a small number of lesions can be difficult because small areas tend to look healthy. It directly reduces the classification accuracy, and extracting and choosing informative features is challenging. A significant role is played by automatically classifying early-stage brain tumors utilizing deep and machine learning approaches. This paper proposes a hybrid deep learning model Convolutional Neural Network-Long Short Term Memory (CNN-LSTM) for classifying and predicting brain tumors through Magnetic Resonance Images (MRI). We experiment on an MRI brain image dataset. First, the data is preprocessed efficiently, and then, the Convolutional Neural Network (CNN) is applied to extract the significant features from images. The proposed model predicts the brain tumor with a significant classification accuracy of 99.1%, a precision of 98.8%, recall of 98.9%, and F1-measure of 99.0%.

1. Introduction

The era of crucial medical data is approaching as a result of the quick advancements in medical technology. The monitoring, prevention, and diagnosis of infectious tumors depend on the management, collection, and appropriate analysis of data in infectious tumor diagnosis and treatment (Akram et al., 2021; Houssein et al., 2021a; Zhang et al., 2021; Javed et al., 2022; Rizwan et al., 2022b). Cancer is one of the most significant health issues and challenges currently threatening people's lives. The scariest and most dangerous sort of cancer among the various cancers is brain tumors (Ali et al., 2022; Senan et al., 2022). The brain is the most complex organ of the human body that acts through billions of cells and connections called synapses (Alanazi et al., 2022). The human brain is the nervous control center system to control all organs of the human body (Mewada et al., 2020; Rizwan et al., 2022a). Therefore, having an abnormal brain has disastrous effects on people's health. The World Health Organization (WHO) claims that, in 2020, roughly 10 million deaths were recorded because of brain cancer, which is the second-leading death cause globally (Can, 2022). Cancer is considered the most deadly and devastating disease because of its diverse characteristics, less survival rate, and aggressive nature. A misdiagnosed brain tumor results in ineffective medical treatment, which lowers the chances of survival for the patient (Rathod and Khan, 2021; Senan et al., 2022). Different types of tumors are based on their texture, location, and shape, such as the pituitary, glioma, lymphomas, Medulloblastoma, malignant and acoustic neuroma (Alanazi et al., 2022; Kumar et al., 2022).

The major challenge arises in detecting the brain tumor because of the tumor location, type, size, and shape variations (Amin et al., 2021). The brain tumor diagnosis depends on the type and location of the tumor so that the doctors can predict the survival chances of the patients and make decisions for treatment that extent from surgery to radiotherapy and chemotherapy (Kumar et al., 2022). Therefore, identifying and detecting the brain tumor at early stages helps plan the treatment and monitor the condition of patients. It plays a crucial part in enhancing the treatment and ensuring higher chances of survival. Different types of medical imaging and diagnostic techniques are used to obtain information related to tumors. Magnetic Resonance Imaging (MRI) and Computed Tomography (CT) are methods for identifying normal and abnormal growing cells in the brain (Houssein et al., 2021b; Alsaif et al., 2022). The Computed Tomography (CT) scan is used to diagnose the patient using an X-ray and the computer to create the brain images in the axial fragments (Bekhet et al., 2020; Fahmi et al., 2020; Mubashar et al., 2022).

Magnetic resonance imaging creates 2D and 3D images of the body's internal organs without pain or needing surgery (Alanazi et al., 2022). It is one of the most accurate methods for the early identification and diagnosis of cancer. However, determining the type of tumor with MRI is challenging, error-prone, and time-consuming, necessitating radiologists with extensive experience. Due to the variety of tumors, MRI images occasionally have no apparent features that would enable sound decision-making. Consequently the human cannot depends on manual diagnosis (Al-Shoukry et al., 2020; Tiwari et al., 2022). An accurate diagnosis helps the patient to receive adequate treatment and longtime survival. Furthermore, the brain tumor's under-diagnosis is risky because it lowers the effectiveness of treatment and chances of survival. Consequently, using artificial intelligence (AI) techniques has become necessary in the computer-aided diagnostic (CAD) system's ability to diagnose medical images like MRI scans. These techniques help doctors and radiologists make an accurate diagnosis while reducing burden (Al-Shoukry et al., 2020; Gab Allah et al., 2021). The CAD system is divided into different steps, such as a preprocessing step for eliminating the noise from images, a segmentation step used to identify the lesion area and separate it from the rest of the image, a feature extraction step for extracting the significant features that show the tumor and a classification step to classify each image and predicts the abnormality (Lu et al., 2021).

Motivation: The ability to distinguish between abnormal and normal tissues makes detecting brain tumors challenging. Brain tumors should be found as soon as possible to allow the patient to recover quickly from the treatment. Patients can receive better care if the brain tumor is initially detected. Early brain tumor detection allows for more incredible characterization of the brain's system structures and their easy identification (Kumar et al., 2022). Despite extensive research, there are still certain limitations in detecting brain tumors because of the unusual distribution pattern of the lesions. Finding a region with a small number of lesions can be difficult because small areas tend to look healthy. Extracting and selecting informative features is also tricky because it decreases classification accuracy (Amin et al., 2022). According to the literature review, numerous machine learning techniques have been employed to categorize MRI images (Rehman et al., 2020; Zhou et al., 2020; Kumar et al., 2021). Recent advances in machine learning have led to the application of numerous deep learning approaches for diagnosing MRI images (Alanazi et al., 2022; Alrashedy et al., 2022; Qureshi et al., 2022; Senan et al., 2022; Zeineldin et al., 2022). The main contributions of the research are given below:

• Propose a hybrid deep learning model based on Convolutional Neural Network-Long Short Term Memory (CNN-LSTM) for classifying and predicting brain tumors through Magnetic Resonance Images (MRI).

• Convolutional neural network removes the noise and extracts the essential features.

• The hybrid deep learning model handles the large capacity of high-dimensional data and improves the model's performance.

• Proposed model provides a promising diagnostic model to diagnose the MRI brain images for brain tumor classification and support the radiologists' and experts' decisions.

The paper is organized into the following sections: Section 2 discusses the literature review of the machine learning and deep learning approaches for identifying brain tumors. Section 3 describes the research design of the proposed work, using the MR image dataset and deep learning models. Section 4 explains the results and discusses them. Sections 5 present the work's conclusion and recommendations for future research.

2. Literature review

Many studies are conducted to find brain tumors, and some of the most recent methods are mentioned in this section.

2.1. Machine learning techniques

Different machine learning techniques are implemented for various healthcare applications such cognitive health assessment, cervical cancer detection, tumor detection, breast cancer detection (Javed et al., 2020, 2021a,b; Mehmood et al., 2021; Mohiyuddin et al., 2022; Rehman et al., 2022b). The machine learning techniques RF, SVM, AdaBoost1, and RUSBoost, are used in Rehman et al. (2020) to localize the brain tumor on FLAIR scans MRI. These techniques are implemented on the BraTs 2012 dataset for both natural and syntactic images, and the proposed model indicates the result with the best accuracy is 0.98%, sensitivity is 0.92%, specificity is 0.96%, precision is 0.88%, and the dice score is 0.88%. The automatic brain tumor classification system is proposed in Kumar et al. (2021) and the K-nearest neighbor algorithm is used to classify the MRI images as abnormal or normal. The fuzz C-means clustering technique is used for the segmentation of tumor regions. The two datasets MICCAI and BRATS, are used for the experiment and evaluate the results with 96.5% accuracy, 100% sensitivity, and 93% specificity. The manual optimizing model by a machine learning expert is presented in Zhou et al. (2020) and compared with the automated machine learning technique Tree-Based Pipeline optimizing tool for evaluating the model performance. The proposed model was implemented on MRI images of 288 patients, and results showed that the best AUC value is 0.94% and accuracy is 0.88%.

2.2. Deep learning techniques

Although deep learning models have been employed in many fields, they still need to be adjusted before they can be used in delicate fields like medical imaging. The GAN architecture is proposed in Alrashedy et al. (2022) and different deep learning models CNN, ResNet152V2, and MobileNetV2 are used to generate and categorize the MRI brain images. The images are created by DCGAN and Vanilla GAN, which are used to train the deep transfer models and evaluate the performance on the test set composed of authentic MRI brain images. The experiment results indicate that ResNet152V2 achieved the best results with a 99.09% accuracy, 99.51% AUC, 99.08% recall, 99.12% precision, and the loss is 0.196 based on the MRI brain images. The novel transfer deep learning model is proposed in Alanazi et al. (2022) to diagnose the brain tumor early by using different subclasses such as glioma, pituitary, and meningioma. To assess their performance for the MRI brain pictures, the Convolutional Neural Network models are created from the ground up. The MRI Brain images are then classified into tumor sub-classes using the different 22-layer, binary-classification (tumor or no tumor) isolated-convolutional neural network model by modifying the neurons' weights using the transfer-learning technique. The transfer learning model is implemented on an unseen MRI brain dataset, and the result indicates that the proposed model provides 96.89% high accuracy.

Authors in Amin et al. (2022) the new model is proposed to detect brain cancer using ensemble transfer learning and Quantum Variational classifiers (QVR). The in-depth features are extracted by the inceptionv3 model in which the score vector is obtained by softmax and used the QVR for discrimination among pituitary tumor, no tumor, meningioma, and glioma. The research is implemented on three different datasets such as 2020-BRATS, local images, and Kaggle, and the proposed model achieved more than 90%detection score. The NeuroXAI framework is proposed in Zeineldin et al. (2022) to explain deep learning networks for increasing medical expert trust. The proposed framework implements the seven different methods such as vanilla gradient, guided back-propagation, integrated gradients, guided integrated gradients, Smooth-Grad, Grad-CAM, and guided Grad-CAM for providing the maps of visualization that help in creating the transparent deep learning model. The framework is implemented on the BraTS 2019 dataset and achieved 98.62% accuracy.

The automated Ultra-Light Brain Tumor Detection (UL-BTD) system is proposed in Qureshi et al. (2022) that is based on the new Ultra-Light Learning Architecture (UL-DLA) for the in-depth features, merged with the textural features that extracted from Gray Level Co-occurrence Matrix (GLCM). It created the Hybrid Feature Space (HFC) for detecting the brain tumor using a support vector machine. The proposed system is implemented on a T1-weighted MRI dataset and achieved the 99.23% average detection rate, and 0.99% is the F1 measure. Different experiments are performed in Senan et al. (2022) to diagnose the brain tumor by combining machine learning and deep learning techniques. For identifying and categorizing brain cancers, support vector machine techniques are combined with ResNEt-18 and AlexNet. A brain tumor MRI images are enlarged using the average filter technique, and deep learning techniques are used to extract the essential features using deep convolutional layers. The extracted features are classified by using SVM and Softmax. The experiment is implemented on an MRI dataset containing 3,060 images and divided into four classes: one is normal, and the other three are tumors. The results indicate that AlexNet with SVM showed the best results with a 95.10% accuracy, 98.50% specificity, and 95.25% sensitivity. The detailed review of CNN architectures is presented in Alsaif et al. (2022) and provides the characteristics of different models such as VGG, ResNet, and AlexNet. The method based on CNN and data augmentation is applied to the MRI dataset to detect the brain tumor, and the result showed that the VGG model achieved a high value with a 0.93% accuracy,0.93% F1-score, 0.94% precision, and 0.93% recall.

Some gaps can be utilized to recognize and detect brain tumors, including ensemble classifiers, machine learning methods, and numerous datasets used to test the generalizability of methods like BRATS, T1-weighted MRI dataset, and deep learning model to categorize the MRI brain pictures. Table 1 summarize the existing literature work. It can be noticed that there exists some work on brain tumor detection but the lack in performance.

TABLE 1
www.frontiersin.org

Table 1. Summery of existing literature work.

3. Proposed methodology

In this section, the details of the methods and algorithms of the proposed approach are explained that are used for the classification of MRI Brain images. To validate the proposed approach, accuracy, precision, recall, and F1-measure are calculated as a part of a controlled experiment. Python language is used for conducting a controlled experiment. In an experiment, the impact of the independent variable on the dependent variable is investigated and changed. The accuracy of the suggested model is the dependent variable in this study, while independent variables are employed to examine their effects. The proposed model has four different steps in the first step selects the dataset. In the second step, data pre-processing is performed, which consists of different steps such as thresholding, extreme point calculation, and bicubic interpolation. In the third step, data extraction is performed, and in the last step, the experiment is performed where deep learning technique CNN and CNN-based hybrid deep learning model CNN-LSTM is applied to detect the brain tumor. The proposed technique is presented in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Proposed approach.

3.1. Experimental dataset

The system's performance is assessed on the MRI brain tumor dataset. The dataset is gathered from Kaggle, which is the publically available database. Kaggle allows the user to find and publish different datasets, work with different machine and deep learning publishers and data scientists, and design and explore the models in different data science environment (Dahiwade et al., 2019). The dataset consists of a total number of 253 images. The dataset is arranged into one folder (BrainTumorImages) and consists of two sub-folders, one is no tumor with 98 images, and the other is a tumor folder with 155 images. MRI brain image classification is represented in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. MRI brain images samples for two classes tumor and no tumor.

3.2. Data preprocessing

In this section, data preprocessing is performed to remove the undesired data in noise form that decreases the model's performance. The unwanted areas and spaces are present in every MRI brain dataset image. Therefore, cropping the images is crucial to remove extraneous space and utilize only the relevant data. This research uses the cropping technique in Dahiwade et al. (2019) that uses the extreme point calculation. The steps to crop the MR images using the extreme point calculation are presented in Figure 3. In the first step for preprocessing, load the actual MR image, convert the images into grayscale and blur it slightly and then apply the thresholding to the magnetic resonance image for converting these images into binary images. The erosion and dilation operations are performed to remove any little noise regions in the images. Afterward, select the most significant contour from threshold images and compute four different extreme points (extreme left, extreme right, extreme bottom, and extreme top). In the last step, crop the image using the collected information from extreme points and contours. By using the bicubic interpolation, crop the MR tumor images. Bicubic interpolation is preferred because it can create a smoother curve compared to other interpolation techniques like bilinear interpolation, and it is the best option for MRI brain tumor images since there is more noise at the edge. In the MRI images dataset, the images are of different sizes, heights, and widths. So it is necessary to resize the images into equal height and width to achieve the best results. In this study, the image is resized to 224 X 224 for uniformity. After that, all images are encoded between 0 and 255, and finally, images are normalized.

FIGURE 3
www.frontiersin.org

Figure 3. Data preprocessing steps.

3.3. Feature extraction

After preprocessing the data, the MR image dataset is trained. Features are extracted that represent each tumor. The models based on CNN extract the significant features without any human supervision. The advantage of deep learning decides how to use convolutional filters for extracting the features from the training dataset. In this research, the deep learning models CNN and CNN-LSTM are implemented to classify brain tumor types and extract the in-depth features from an image. The extracted features are then fed to the deep learning models, including a convolutional neural network with a fully connected (FC) layer (Kang et al., 2021).

3.4. Convolutional neural network

Convolutional Neural Network is the deep neural network class that utilizes the different convolutional layers to filter inputs for helpful data. The Convolutional filters are applied to the input through the convolutional layers of CNN to compute the output of the neurons connected to specific regions in the input. It helps to extract the temporal and spatial features from an image (Goyal et al., 2019).

The CNN model contains three significant layers a convolutional layer, a max-pooling layer, and a fully connected layer. The convolutional layer contains three significant parameters: pitch, padding, and filter size. Many filters in each layer are used for in-depth feature extraction. According to stride, the filters move within the images. The stride size is either one or two; if the value is more significant than two, CNN performance deteriorates. The zero padding is necessary for retaining the structural assessment when the filter does not shield all of the input images in the convolutional layer. The goal of each convolutional layer is to perform a particular task; for example, the edges of the lesions are highlighted in the first layer, the second layer is used to extract the features of complex geometrics, and in the third layer, the lesions shapes and colors are highlighted. In the feature map, the ReLU layer passes the positive values while negative values are suppressed, turning them into zero (Senan et al., 2022). The max-pooling layer is used to downsample or decrease the dimensionality of the extracted features. The max and average are the two popular methods of the max-pooling layer. The fully connected layer with the 512 unit is used to classify the image into different classes (Goyal et al., 2019; Kang et al., 2021). For feature map normalization, the batch normalization layer is used. These layers accelerate the training and network regulation. The dropout layer is employed in some issues and is highly beneficial to overcome the over-fitting issues with networks (Balaha et al., 2021). The CNN architecture is presented in Figure 4.

FIGURE 4
www.frontiersin.org

Figure 4. Architecture of convolutional neural network.

3.5. Long short term memory

The long short-term memory is the improved version of the Recurrent Neural Network (RNN). Long-term memory suggests the memory blocks rather than traditional units of RNN to solve the problem of exploding gradient and vanishing (Amin et al., 2020). The main difference between LSTM and RNN is that LSTM adds the cell state for saving the long-term states. An LSTM network can recall and link data from the past to the information acquired in the present (Chen, 2016). The long short-term memory is designed with three different gates, the input gate, forget gate, and the output gate. The current input is represented by ut the new and the previous states of the cell are represented by st and st−1 respectively, and the current and the previous output is denoted by vt and vt−1 respectively. The rules for the input gate of the LSTM are given in the below equations.

In Equation (1) pass the previous output vt−1 and ut by the sigmoid layer to determine which part of the information has to be added by using the Jt to take the input information while ai represents the input gate.

Jt=σ(Wi.[vt-1,ut]+ai)    (1)

The Equation (2) is used to get the updated information St after passing the previous information vt−1 and current information ut by tanh layer using the input gate ai. The current state of information st, and information of long-term memory St−1 into St are integrated in Equation (3), where sigmoid output is referred by Wi and Stis referred the tanh output. The weight metrics are represented by Wi, and the input gate of the LSTM is represented by ai. The forget gate of the LSTM then permits the selective transmission of the information by using the dot product of input information Jt and the current state of information st and sigmoid layer.

st=tanh(Wi.[vt-1,ut]+ai)    (2)
St=ht(St-1)+jtst    (3)

With the specific probability, it is decided whether to delete the related data by the last cell by using Equation (4), where Wf represents the weight matrix, offset is af and sigmoid function is σ.

bt=σ(Wf.[vt-1,ut]+af)    (4)

The inputs in Equations (5) and (6) determine that the output gate Ot of the LSTM describes the states that are needed for the continuation through the previous and current output vt−1 and ut respectively. The final output Vt is multiplied and acquired through the decision vector of the state that transfers the new data St by the tanh layer.

Ot=σ(Wo.[vt-1,ut]+ao)    (5)
Vt=Ottanh(St)    (6)

Where the Wo is the weighted matrix of the output gate ao is the bias of LSTM of output gate the (Islam et al., 2020).

The architecture of LSTM model is presented in Figure 5. The model has three gates: the input, forget, and output. The first gate is forget gate (ht) that takes the previous output vt−1 and current input ut form previous state st−1 with sigmoid function σ. The input gate add the previous information and take input information Jt by tanh layer and sigmoid function σ. To get the output, the output gate is used where the information comes from the input gate, and the output gate computes all information by tanh layer and uses the sigmoid function σ and provides the current state where the output is stored.

FIGURE 5
www.frontiersin.org

Figure 5. Architecture of LSTM model.

3.6. Proposed hybrid CNN-LSTM model

In this research, the combined model is developed to detect the brain tumor automatically. The architecture's structure is created by combining the CNN and networks of LSTM, where CNN extracts the features, and as a classifier, LSTM is used.

Figure 6 presents the proposed hybrid model for brain tumor detection. The network has 16 layers, three convolutional layers, seven max-pooling layers, four ConvLSTM layers, one fully connected layer, and the last one is the output layer connected with the softmax function. The proposed model first uses the three convolutional layers as CNN layers with three max-pooling layers and then uses the four ConvLSTM with four max-pooling layers. Each ConvLSTM and the convolutional block are integrated with a 2D convolutional neural network, one max-pooling layer, and a dropout layer with a 25% rate. The convolutional layer with a kernel size of 3 x 3 is used to extract the features from MRI brain images activated through the function of Relu. The max-pooling layer with kernel size 2 x 2 reduces the input image dimensionality. The convolutional layer and max-pooling layers are used with different input sizes such as (64,64,3), (56, 56,128), (112, 112, 128), (112, 112, 64) and (224, 224, 128) for less shrinking of the images. Fewer shrinkage results in less distortion of the image's internal characteristics and patterns. After the convolutional block, the output shape is found as the (56, 56, 128) given in the ellipsis form in Figure 6. Then using the reshaping method to resize the images into equal height and width to achieve the best results, after the reshaping method, the size of the ConvLSTM layer becomes (56,512) used for sequence prediction. After reshaping the ConvLSTM layer, the size of the fully connected layer and the output layer has 512 and 64, respectively. In the architecture's last part, The ConvLSTM layer is used for extracting the information about time. The architecture categorizes the MRI brain tumor image, either tumor or non-tumor, by a fully connected layer after looking at the characteristics of the time.

FIGURE 6
www.frontiersin.org

Figure 6. Architecture of proposed hybrid CNN-LSTM model.

In Table 2 summary of the proposed architecture CNN-LSTM is provided. The layers types, kernel size, and input size are provided in detail. The convolutional2D layer has the 3 x 3 size and different input sizes such as (64 x64 x 3) and (56 x 56 x 128). The maxpooling2D layer has the 2 x 2 kernel size with (64 x 64 x 3), (56 x 56 x 128), (112 x 112 x 64), (112 x 112 x 128) and (14 x 14 x 128) input size. The ConvLSTM2D layer has kernel size 3 x 3 with different input sizes such as (112 x 112 x 64), (112 x 112 x 128) and (224 x 224 x 128). The fully connected layer and the output layer have the input sizes 512 and 64, respectively. The convolutional and max-pooling layers are employed for reduced image compression with varied input sizes. Less contraction leads to less distortion of the internal traits and patterns of the image. By a fully connected layer, the architecture sorts the MRI brain tumor image, either tumor or not tumor, after analyzing the characteristics of the time. The Adam optimizer is used for training the proposed model with a batch size of 32, the number of epochs is 100, and cross-entropy is used for losses and the metrics for accuracy.

TABLE 2
www.frontiersin.org

Table 2. The CNN-LSTM network summary.

The steps in the proposed model CNN-LSTM are given in Algorithm 1. The image from the MR brain tumor dataset is taken as input. Image preprocessing is performed by resizing the image to a size of 224 x 224 and cropping the images through the extreme point calculation method. Split the data into the parts of training (70%) and validation (30%). After this, classify and extract the features through a convolutional neural network and hybrid deep learning model CNN-LSTM with 100 epochs and 32 batch sizes. Calculate the loss through cross entropy and apply optimization and fitness functions to validate and train the model. Calculate the performance metrics and obtain the result.

ALGORITHM 1
www.frontiersin.org

Algorithm 1. Pseudo code of Proposed CNN-LSTM model

4. Result and discussion

In this section, analyze the proposed model performance. The proposed model is evaluated on different parameters. These parameters are carried out to determine whether the proposed model is better than the previous methods and whether it is appropriate for brain tumor detection or not. The proposed model is implemented on an MRI brain image dataset, and a deep learning model is applied to this dataset.

4.1. Evaluation metrics

Different evaluation metrics are used for the prediction and classification challenges, such as accuracy, precision, recall, and F1-measure (Rehman et al., 2022a). The effectiveness of the proposed model is assessed using the following evaluation metrics.

Accuracy: To measure the accuracy of the proposed model, compute the ratio of the false positive, true positive, true negative, and false negative. The Equation (7) represent the accuracy estimate.

Accuracy=TP+TNTP+TN+FP+FN    (7)

Precision: The ratio of real positives to all positives overall (both false and true) in the data. Additionally referred to as a high predicted value. The precision rate is represented in Equation (8).

Precision=TPTP+FP    (8)

Recall: The ratio of the true positives in the data to true positives and false negatives is often referred to as sensitivity, the chance of detection, and the rate of a true positive. The recall rate is represented in Equation (9).

Recall=TPTN+FN    (9)

F1-measure: The weighted average of the precision and recall is known as the F1 measure. The Equation (10) represents the value of F1-measure.

F1-measure=2×Precision+RecallPrecision+Recall    (10)

4.2. Experimental analysis and result

The proposed technique is applied to deep learning models CNN and CNN-LSTM. Figure 7 represents the performance evaluation of the convolutional neural network with training and validation loss and accuracy. At every 100 epochs, the training accuracy is 99.4%, and validation accuracy is 98.3%. Similarly, the training loss is 0.007, and 0.113 is the validation loss.

FIGURE 7
www.frontiersin.org

Figure 7. Evaluation metrics of CNN architecture.

The performance evaluation of the CNN-LSTM with training and validation loss and accuracy is represented in Figure 8. At every 100 epochs, the training and validation accuracy is 99.8 and 98.5%, respectively. Similarly, the training loss is 0.010, and 0.103 is the validation loss.

FIGURE 8
www.frontiersin.org

Figure 8. Evaluation metrics of CNN-LSTM architecture.

Table 3 provides the results of proposed technique. The CNN achieved the accuracy is 98.6%, the precision is 98.5%, a recall is 98.6%, and F1-measure is 98.4%. The hybrid deep learning model CNN-LSTM achieved an accuracy of 99.1%, a precision of 98.8%, a recall of 98.9%, and an F1-measure of 99.0%.

TABLE 3
www.frontiersin.org

Table 3. Performance of CNN and Hybrid CNN-LSTM model.

Figure 9 graphically represents the performance of the proposed model. The proposed model CNN-LSTM performs well compared to the deep learning technique CNN in terms of accuracy 99.1%, precision 98.8%, recall 98.8%, and F1-measure 99.0%.

FIGURE 9
www.frontiersin.org

Figure 9. Performance of proposed model.

4.3. Comparative analysis

The comparison of the proposed model with previous techniques is provided in Table 4. The same preprocessing steps are performed in the previous techniques as the preprocessing steps for the proposed model.

TABLE 4
www.frontiersin.org

Table 4. Comparison with previous techniques.

Authors in Alsaif et al. (2022) provides the result in term of accuracy 93%, precision 94%, recall 93% and F1-measure 93%. Authors in Rehman et al. (2020) achieved the result in terms of accuracy 98% and precision 88%. Authors in Salama and Shokry (2022) achieved the results in terms of accuracy 97%, precision 96.88%, recall 96%, and F1-measure of 96.88%.

In Figure 10 the comparison of the proposed model with the existing one is graphically represented. It seems that the proposed model is performing better than previous techniques.

FIGURE 10
www.frontiersin.org

Figure 10. Comparison of proposed model with existing techniques.

5. Conclusion

Detecting a brain tumor is complicated because of the brain's complex structure. Every organ in the body has a function that is controlled by the brain. Automatic initial stage brain tumor categorization using deep learning and machine learning techniques plays a crucial part. These systems enable prompt diagnosis and raise the likelihood of survival for patients. Additionally, these methods support specialists and radiologists in their decision-making regarding diagnoses and plans for treatment. This study proposed the CNN-based hybrid deep learning model CNN-LSTM to classify the brain tumors using the MR brain tumor images dataset; firstly, the image dataset is by thresholding, extreme point calculation, and bicubic interpolation. Secondly, the proposed model uses the convolutional neural network for extracting the features in the form of cropped images. Four metrics, accuracy, precision, recall, and F1-measure, are used to evaluate the model's performance. The proposed model provides the best result by achieving 99.1% accuracy, precision is 98.8%, recall is 98.9%, and F1-measure is 99.0%. The results showed that the proposed model is best for detecting the MR brain images. The future work would be to investigate the performance of the proposed approach on multi-class MR brain tumor images problem and use different datasets such as Brast2022 and T-weighted to enhance the performance of the proposed model.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.

Funding

The research is supported by Qatar National Library and Qatar university internal grant IRCC-2021-010.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Akram, F., Liu, D., Zhao, P., Kryvinska, N., Abbas, S., and Rizwan, M. (2021). Trustworthy intrusion detection in e-healthcare systems. Front. Public Health 9, 788347. doi: 10.3389/fpubh.2021.788347

PubMed Abstract | CrossRef Full Text | Google Scholar

Alanazi, M. F., Ali, M. U., Hussain, S. J., Zafar, A., Mohatram, M., Irfan, M., et al. (2022). Brain tumor/mass classification framework using magnetic-resonance-imaging-based isolated and developed transfer deep-learning model. Sensors 22, 372. doi: 10.3390/s22010372

PubMed Abstract | CrossRef Full Text | Google Scholar

Ali, T. M., Nawaz, A., Rehman, A. U., Ahmad, R. Z., Javed, A. R., Gadekallu, T. R., et al. (2022). A sequential machine learning-cum-attention mechanism for effective segmentation of brain tumor. Front Oncol. 12, 873268. doi: 10.3389/fonc.2022.873268

PubMed Abstract | CrossRef Full Text | Google Scholar

Alrashedy, H. H. N., Almansour, A. F., Ibrahim, D. M., and Hammoudeh, M. A. A. (2022). Braingan: brain mri image generation and classification framework using gan architectures and cnn models. Sensors 22, 4297. doi: 10.3390/s22114297

PubMed Abstract | CrossRef Full Text | Google Scholar

Alsaif, H., Guesmi, R., Alshammari, B. M., Hamrouni, T., Guesmi, T., Alzamil, A., et al. (2022). A novel data augmentation-based brain tumor detection using convolutional neural network. Appl. Sci. 12, 3773. doi: 10.3390/app12083773

CrossRef Full Text | Google Scholar

Al-Shoukry, S., Rassem, T. H., and Makbol, N. M. (2020). AlzheimerŠs diseases detection by using deep learning algorithms: a mini-review. IEEE Access 8, 77131–77141. doi: 10.1109/ACCESS.2020.2989396

PubMed Abstract | CrossRef Full Text | Google Scholar

Amin, J., Anjum, M. A., Sharif, M., Jabeen, S., Kadry, S., and Moreno Ger, P. (2022). A new model for brain tumor detection using ensemble transfer learning and quantum variational classifier. Comput. Intell. Neurosci. 2022, 3236305. doi: 10.1155/2022/3236305

PubMed Abstract | CrossRef Full Text | Google Scholar

Amin, J., Sharif, M., Haldorai, A., Yasmin, M., and Nayak, R. S. (2021). Brain tumor detection and classification using machine learning: a comprehensive survey. Complex Intell. Syst. 8, 3161–3183. doi: 10.1007/s40747-021-00563-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Amin, J., Sharif, M., Raza, M., Saba, T., Sial, R., and Shad, S. A. (2020). Brain tumor detection: a long short-term memory (lstm)-based learning model. Neural Comput. Appl. 32, 15965–15973. doi: 10.1007/s00521-019-04650-7

CrossRef Full Text | Google Scholar

Balaha, H. M., El-Gendy, E. M., and Saafan, M. M. (2021). Covh2sd: A COVID-19 detection approach based on harris hawks optimization and stacked deep learning. Expert. Syst. Appl. 186, 115805. doi: 10.1016/j.eswa.2021.115805

PubMed Abstract | CrossRef Full Text | Google Scholar

Bekhet, S., Hassaballah, M., Kenk, M. A., and Hameed, M. A. (2020). “An artificial intelligence based technique for COVID-19 diagnosis from chest x-ray,” in 2020 2nd Novel Intelligent and Leading Emerging Sciences Conference (NILES) (Giza: IEEE), 191–195.

Google Scholar

Can (2022). Cancer. Available online at: https://www.who.int/news-room/fact-sheets/detail/cancer (accessed on 30 June, 2022).

Chen, G. (2016). A gentle tutorial of recurrent neural network with error backpropagation. arXiv preprint arXiv:1610.02583. doi: 10.48550/arXiv.1610.02583

CrossRef Full Text | Google Scholar

Dahiwade, D., Patle, G., and Meshram, E. (2019). “Designing disease prediction model using machine learning approach,” in 2019 3rd International Conference on Computing Methodologies and Communication (ICCMC) (Erode: IEEE), 1211–1215.

Google Scholar

Fahmi, F., Apriyulida, F., Nasution, I. K., et al. (2020). Automatic detection of brain tumor on computed tomography images for patients in the intensive care unit. J. Healthc Eng. 2020, 2483285. doi: 10.1155/2020/2483285

PubMed Abstract | CrossRef Full Text | Google Scholar

Gab Allah, A. M., Sarhan, A. M., and Elshennawy, N. M. (2021). Classification of brain mri tumor images based on deep learning pggan augmentation. Diagnostics 11, 2343. doi: 10.3390/diagnostics11122343

PubMed Abstract | CrossRef Full Text | Google Scholar

Goyal, M., Goyal, R., and Lall, B. (2019). Learning activation functions: A new paradigm for understanding neural networks. arXiv preprint arXiv:1906.09529. doi: 10.1007/978-3-030-31760-7_1

CrossRef Full Text | Google Scholar

Houssein, E. H., Ibrahim, I. E., Hassaballah, M., and Wazery, Y. M. (2021a). “Integration of internet of things and cloud computing for cardiac health recognition,” in Metaheuristics in Machine Learning: Theory and Applications (Springer), 645–661.

Google Scholar

Houssein, E. H., Ibrahim, I. E., Neggaz, N., Hassaballah, M., and Wazery, Y. M. (2021b). An efficient ecg arrhythmia classification method based on manta ray foraging optimization. Expert. Syst. Appl. 181, 115131. doi: 10.1016/j.eswa.2021.115131

CrossRef Full Text | Google Scholar

Islam, M. Z., Islam, M. M., and Asraf, A. (2020). A combined deep cnn-lstm network for the detection of novel coronavirus (COVID-19) using x-ray images. Inform. Med. Unlocked 20, 100412. doi: 10.1016/j.imu.2020.100412

PubMed Abstract | CrossRef Full Text | Google Scholar

Javed, A. R., Fahad, L. G., Farhan, A. A., Abbas, S., Srivastava, G., Parizi, R. M., et al. (2021a). Automated cognitive health assessment in smart homes using machine learning. Sustain. Cities Soc. 65, 102572. doi: 10.1016/j.scs.2020.102572

CrossRef Full Text | Google Scholar

Javed, A. R., Sarwar, M. U., Beg, M. O., Asim, M., Baker, T., and Tawfik, H. (2020). A collaborative healthcare framework for shared healthcare plan with ambient intelligence. Human Centric Comput. Inf. Sci. 10, 1–21. doi: 10.1186/s13673-020-00245-7

CrossRef Full Text | Google Scholar

Javed, A. R., Sarwar, M. U., Khan, H. U., Al-Otaibi, Y. D., Alnumay, W. S., et al. (2021b). Pp-spa: privacy preserved smartphone-based personal assistant to improve routine life functioning of cognitive impaired individuals. Neural Process. Lett. 1–18. doi: 10.1007/s11063-020-10414-5

CrossRef Full Text | Google Scholar

Javed, A. R., Shahzad, F., ur Rehman, S., Zikria, Y. B., Razzak, I., Jalil, Z., et al. (2022). Future smart cities requirements, emerging technologies, applications, challenges, and future aspects. Cities 129, 103794. doi: 10.1016/j.cities.2022.103794

PubMed Abstract | CrossRef Full Text | Google Scholar

Kang, J., Ullah, Z., and Gwak, J. (2021). Mri-based brain tumor classification using ensemble of deep features and machine learning classifiers. Sensors 21, 2222. doi: 10.3390/s21062222

PubMed Abstract | CrossRef Full Text | Google Scholar

Kumar, D. M., Satyanarayana, D., and Prasad, M. (2021). Mri brain tumor detection using optimal possibilistic fuzzy c-means clustering algorithm and adaptive k-nearest neighbor classifier. J. Ambient. Intell. Humaniz Comput. 12, 2867–2880. doi: 10.1007/s12652-020-02444-7

CrossRef Full Text | Google Scholar

Kumar, T. S., Arun, C., and Ezhumalai, P. (2022). An approach for brain tumor detection using optimal feature selection and optimized deep belief network. Biomed. Signal Process. Control 73, 103440. doi: 10.1016/j.bspc.2021.103440

CrossRef Full Text | Google Scholar

Lu, S.-Y., Satapathy, S. C., Wang, S.-H., and Zhang, Y.-D. (2021). PBTNet: a new computer-aided diagnosis system for detecting primary brain tumors. Front. Cell Dev. Biol. 9, 765654. doi: 10.3389/fcell.2021.765654

PubMed Abstract | CrossRef Full Text | Google Scholar

Mehmood, M., Rizwan, M., Abbas, S., and Abbas, S. (2021). Machine learning assisted cervical cancer detection. Front. Public Health 9, 788376. doi: 10.3389/fpubh.2021.788376

PubMed Abstract | CrossRef Full Text | Google Scholar

Mewada, H. K., Patel, A. V., Hassaballah, M., Alkinani, M. H., and Mahant, K. (2020). Spectral-spatial features integrated convolution neural network for breast cancer classification. Sensors 20, 4747. doi: 10.3390/s20174747

PubMed Abstract | CrossRef Full Text | Google Scholar

Mohiyuddin, A., Basharat, A., Ghani, U., Abbas, S., Naeem, O. B., and Rizwan, M. (2022). Breast tumor detection and classification in mammogram images using modified yolov5 network. Comput. Math. Methods Med. 2022, 1359019. doi: 10.1155/2022/1359019

PubMed Abstract | CrossRef Full Text | Google Scholar

Mubashar, A., Asghar, K., Javed, A. R., Rizwan, M., Srivastava, G., Gadekallu, T. R., et al. (2022). Storage and proximity management for centralized personal health records using an ipfs-based optimization algorithm. J. Circ. Syst. Comput. 31, 2250010. doi: 10.1142/S0218126622500104

CrossRef Full Text | Google Scholar

Qureshi, S. A., Raza, S. E. A., Hussain, L., Malibari, A. A., Nour, M. K., Rehman, A., et al. (2022). Intelligent ultra-light deep learning model for multi-class brain tumor detection. Appl. Sci. 12, 3715. doi: 10.3390/app12083715

CrossRef Full Text | Google Scholar

Rathod, R., and Khan, R. A. H. (2021). Brain tumor detection using deep neural network and machine learning algorithm. PalArchs J. Archaeol. Egypt Egyptol. 18, 1085–1093. Available online at: https://archives.palarch.nl/index.php/jae/article/view/8814

Google Scholar

Rehman, A., Razzak, I., and Xu, G. (2022a). Federated learning for privacy preservation of healthcare data from smartphone-based side-channel attacks. IEEE J. Biomed. Health Informa. doi: 10.1109/JBHI.2022.3171852

PubMed Abstract | CrossRef Full Text | Google Scholar

Rehman, S. U., Khan, M. U., Nazar Awan, M., Farukh, A., and Hussien, A. (2022b). Personalisedcomfort: a personalised thermal comfort model to predict thermal sensation votes for smart building residents. Enterprise Inf. Syst. 16, 1852316. doi: 10.1080/17517575.2020.1852316

CrossRef Full Text | Google Scholar

Rehman, Z. U., Zia, M. S., Bojja, G. R., Yaqub, M., Jinchao, F., and Arshid, K. (2020). Texture based localization of a brain tumor from mr-images by using a machine learning approach. Med. Hypotheses 141, 109705. doi: 10.1016/j.mehy.2020.109705

PubMed Abstract | CrossRef Full Text | Google Scholar

Rizwan, M., Shabbir, A., Javed, A. R., Shabbir, M., Baker, T., and Obe, D. A.-J. (2022a). Brain tumor and glioma grade classification using gaussian convolutional neural network. IEEE Access 10, 29731–29740. doi: 10.1109/ACCESS.2022.3153108

CrossRef Full Text | Google Scholar

Rizwan, M., Shabbir, A., Srivastava, G., Gadekallu, T. R., Shabir, M., and Hassan, M. A. (2022b). Risk monitoring strategy for confidentiality of healthcare information. Comput. Electr. Eng. 100, 107833. doi: 10.1016/j.compeleceng.2022.107833

CrossRef Full Text | Google Scholar

Salama, W. M., and Shokry, A. (2022). A novel framework for brain tumor detection based on convolutional variational generative models. Multimed. Tools Appl. 81, 16441–16454. doi: 10.1007/s11042-022-12362-9

CrossRef Full Text | Google Scholar

Senan, E. M., Jadhav, M. E., Rassem, T. H., Aljaloud, A. S., Mohammed, B. A., and Al-Mekhlafi, Z. G. (2022). Early diagnosis of brain tumour mri images using hybrid techniques between deep and machine learning. Comput. Math. Methods Med. 2022, 8330833. doi: 10.1155/2022/8330833

PubMed Abstract | CrossRef Full Text | Google Scholar

Tiwari, P., Pant, B., Elarabawy, M. M., Abd-Elnaby, M., Mohd, N., Dhiman, G., et al. (2022). Cnn based multiclass brain tumor detection using medical imaging. Comput. Intell. Neurosci. 2022, 1830010. doi: 10.1155/2022/1830010

PubMed Abstract | CrossRef Full Text | Google Scholar

Zeineldin, R. A., Karar, M. E., Elshaer, Z., Wirtz, C. R., Burgert, O., Mathis-Ullrich, F., et al. (2022). “Explainability of deep neural networks for mri analysis of brain tumors,” in International Journal of Computer Assisted Radiology and Surgery.

PubMed Abstract | Google Scholar

Zhang, L., Zhong, Q., and Yu, Z. (2021). Optimization of tumor disease monitoring in medical big data environment based on high-order simulated annealing neural network algorithm. Comput. Intell. Neurosci. 2021, 8996673. doi: 10.1155/2021/8996673

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, H., Hu, R., Tang, O., Hu, C., Tang, L., Chang, K., et al. (2020). Automatic machine learning to differentiate pediatric posterior fossa tumors on routine mr imaging. Am. J. Neuroradiol. 41, 1279–1285. doi: 10.3174/ajnr.A6621

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: brain tumor, convolutional neural network, long short-term memory, CNN-LSTM, MR images, deep learning

Citation: Alsubai S, Khan HU, Alqahtani A, Sha M, Abbas S and Mohammad UG (2022) Ensemble deep learning for brain tumor detection. Front. Comput. Neurosci. 16:1005617. doi: 10.3389/fncom.2022.1005617

Received: 28 July 2022; Accepted: 18 August 2022;
Published: 02 September 2022.

Edited by:

M. Hassaballah, South Valley University, Egypt

Reviewed by:

Saurabh Singh, Dongguk University Seoul, South Korea
Prabhat Kumar, Lappeenranta University of Technology, Finland

Copyright © 2022 Alsubai, Khan, Alqahtani, Sha, Abbas and Mohammad. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Habib Ullah Khan, habib.khan@qu.edu.qa; Sidra Abbas, sidra.abbas708@gmail.com; Mohemmed Sha, ms.mohamed@psau.edu.sa

Download