Brain tumor classification: a novel approach integrating GLCM, LBP and composite features

Identifying and classifying tumors are critical in-patient care and treatment planning within the medical domain. Nevertheless, the conventional approach of manually examining tumor images is characterized by its lengthy duration and subjective nature. In response to this challenge, a novel method is proposed that integrates the capabilities of Gray-Level Co-Occurrence Matrix (GLCM) features and Local Binary Pattern (LBP) features to conduct a quantitative analysis of tumor images (Glioma, Meningioma, Pituitary Tumor). The key contribution of this study pertains to the development of interaction features, which are obtained through the outer product of the GLCM and LBP feature vectors. The utilization of this approach greatly enhances the discriminative capability of the extracted features. Furthermore, the methodology incorporates aggregated, statistical, and non-linear features in addition to the interaction features. The GLCM feature vectors are utilized to compute these values, encompassing a range of statistical characteristics and effectively modifying the feature space. The effectiveness of this methodology has been demonstrated on image datasets that include tumors. Integrating GLCM (Gray-Level Co-occurrence Matrix) and LBP (Local Binary Patterns) features offers a comprehensive representation of texture characteristics, enhancing tumor detection and classification precision. The introduced interaction features, a distinctive element of this methodology, provide enhanced discriminative capability, resulting in improved performance. Incorporating aggregated, statistical, and non-linear features enables a more precise representation of crucial tumor image characteristics. When utilized with a linear support vector machine classifier, the approach showcases a better accuracy rate of 99.84%, highlighting its efficacy and promising prospects. The proposed improvement in feature extraction techniques for brain tumor classification has the potential to enhance the precision of medical image processing significantly. The methodology exhibits substantial potential in facilitating clinicians to provide more accurate diagnoses and treatments for brain tumors in forthcoming times.


Introduction
A tumor is an abnormal growth of cells which occurs in any portion of the human body.Over two hundred various kinds of cancer, such as lung, blood, breast, heart, lymphoma, etc. have been reported (1).According to World Health Organization (WHO) fact sheet 2022, cancer has been the leading cause of death with 10 million deaths reported (2).Among the various types of tumors, brain tumors have been the primary reason for death in various age and gender groups and are also challenging to treat.A tumor in the human brain is a collection of malignant cells which develops when brain tissues suddenly and abnormally extend.There are different types of brain tumors, some are non-cancerous (benign) and some are cancerous (3).The human brain acts as the body's control hub.It coordinates the actions of vast numbers of neurons and their many connections.Tumor in the brain disrupts normal brain activities and the nervous system processes.The need to overcome the disadvantages of manual tumor image analysis, which is both timeconsuming and vulnerable to human subjectivity, motivated machine learning based techniques of classifying tumors.
As discussed by Abdusalomov et al. ( 4) Glioma, Meningioma, Pitutary seem to be the common types of brain tumors that look like non-cancerous, but may be.Hence this research intends to study these brain tumors and classify them by incorporating advanced features such as GLCM and LBP, as well as interaction features and statistical analysis.This method has the potential to significantly improve the precision of medical image processing for more precise brain tumor identification and treatment planning.
The primary contributions of this present investigation are • This research introduces a novel methodology for comprehensive texture analysis of tumor images.The approach integrates Gray-Level Co-occurrence Matrix (GLCM) and Local Binary Pattern (LBP) features.GLCM features capture spatial pixel intensity relationships, while LBP features identify local texture patterns.Together, these features provide a detailed insight into tumor textures, facilitating improved understanding and tumor classification.• A novel feature generation technique has been proposed, which generates interaction features by multiplying GLCM and LBP feature vectors.This technique generates a new set of features that enhance the discriminative ability of the model, thereby enhancing its capacity to differentiate between various tumor morphologies.The incorporation of these interaction features enables the acquisition of a broader spectrum of texture data, resulting in a greater comprehension of tumor characteristics.• In addition, the methodology incorporates the computation of aggregated characteristics derived from GLCM properties.These aggregated features, which consist of the sum, mean, and median of the GLCM features, provide a more comprehensive view of the overall characteristics of tumor textures.• For Tumor classification Support Vector Machine (SVM) is implemented with extracted features which enhances the performance of tumor classification.
The next section discusses the various research works related to tumors and their classification.

Related literature
In order to categorize common brain tumor types, Kaplan et al. (5) have used nLBP and aLBP feature extraction approaches.Using the K-Nearest Neighbour (Knn) model and the nLBPd = 1 method, they achieved a high 95.56% success rate of identifying tumors.Another study by Abdusalomov et al. (4) used YOLOv7 and transfer learning to improve brain tumor diagnosis in MRI scans, they report an outstanding 99.5% accuracy for identifying the most common types of brain tumors Glioma, Meningioma, Pitutary.However, they also acknowledge the need for additional research, particularly for minor tumor identification (4).
Research by Kaya et al. ( 6) used a novel feature extraction technique based on co-occurrence matrices from vibration data to address the problem of accurate bearing issue identification.Effective success rates were obtained by utilizing 1D-LBP and machine learning: 87.50% for dataset 1 (various speeds), 96.5% for dataset 2 (fault size in mm), and 99.30% for dataset 3 (fault type -inner ring, outer ring, ball).Study by Solani et al. (7) examines the difficulties in diagnosing brain tumors and provides information on the potential of MR imaging.They adopt statistical and machine learning techniques to detect brain tumor for a chosen dataset.Yildirim et al. (8) have studied the most accepted forms of brain tumors include gliomas, meningiomas, and pituitary.They say that the optimal course of action for treating these tumors may differ reliant on the type.Brain tumors can be challenging to classify, even for experts, due to heterogeneous imaging findings (8).
According to Shinde et al. (9), while progress has been made in classifying anomalies in medical imaging, there are still challenges to overcome.These include, but are not limited to, model selection, data description, error detection, data sufficiency, and result reliability.As a result, there is no one highest benchmark for categorizing medical images.So, it is quite problematic in computer vision and machine learning domains.The algorithms mentioned generally are developed using soft computing and model-based methodologies, and their results are reliable (9).With the help of mobile sensor inputs, work by Kuncan et al. (10) presents a unique feature extraction approach called DS-1D-LBP for human activity recognition (HAR).They have successfully classified with the Extreme Learning Machine (ELM) with a high success rate of 96.87%.Research by Shil et al. (11) rely heavily on features that have been manually constructed and then provided to a classifier, such as Support Vector Machine, Decision Tree, or k-Nearest Neighbour (KNN).Khalid et al. state that the extensive nature of the dataset can cause delays in feature engineering, thereby increasing the likelihood of errors and highlighting the significance of domain expertise (12).
Processing and analyzing MRI images of brain tumors is one of the most challenging and promising new areas of study.An MRI, which employs magnetic fields and radio waves to generate overall images of internal body structures, is essential for determining the optimal course of treatment for a tumor and its progression.16) suggested a method for automatically classifying MRI scans for brain tumors using a support vector machine.The researchers employed Fast Fourier Transform (FFT) to extract features to enhance the classifier's precision.Additionally, they utilized technology that exhibited minimal redundancy and maximum relevance to reduce the number of features.
In order to better segregate brain tumors, Shree et al. (17) preprocessed the images using multiple noise removal methods.Their research employed DWT and GLCM-based characteristics of brain tumors.Any residual noise after segmentation was filtered out using morphological filtering procedures.The suggested model was trained and evaluated using the probabilistic neural network classifier for pinpointing tumor locations in brain MRI scans.Yao et al. (18) provided a method that includes extracting texture characteristics using the wavelet transform and classifier as SVM with an accuracy of 83% to process and address protocols of diverse images and non-linearity of actual data to classify improved MRI images related to contrast.Principal component analysis (PCA) and a radial basis function kernel with SVM were proposed by Kumar and Vijayakumar (19) for classifying and segmenting brain tumors.They were able to achieve 94% success rate using this strategy.
Saleck et al. (20) developed a Fuzzy C-Mean (FCM) way of figuring out the size of a patient's brain tumor.They could figure out how many groups were there in the FCM by looking at the intensity of each pixel.This approach uses GLCM texture feature extraction to forecast the threshold value.The generic performance of a model is established by how sensitive, specific, and accurate it is.
Considering the research going on in this field, it seems evident that choosing appropriate features of tumor images and adopting appropriate machine learning classifiers will lead to better identification of tumors.
Based on the literature survey, this research intends to study three types of brain tumors (Glioma, Meningioma, Pitutary) and classify them by incorporating novel and advanced features such as GLCM and LBP, as well as interaction features and statistical analysis.This research work introduces novel contributions in the following aspects: • Interaction Features: The extraction of interaction features involves the computation of the outer product between the Gray-Level Co-occurrence Matrix (GLCM) and Local Binary Pattern (LBP) feature vectors.The aforementioned process results in the formation of a matrix that effectively encompasses the interplay between spatial and local texture data.The matrix is subsequently converted into a unidimensional array in order to constitute the collection of interaction features.• Non-linear Features: Another significant contribution is applying a logarithmic transformation to the GLCM features.This transformation generates non-linear features, allowing for the capturing of complex relationships within tumor images.By incorporating these non-linear features, the research improves the discriminative power of the feature set and enables a more sophisticated analysis and interpretation of the tumor image properties.

Proposed methodology
Publicly available databases Figshare Dataset (13) has been used in this study since it is one of the most common datasets used by many other researchers.This brain tumor dataset contains 3064 T1weighted contrast enhanced images from 233 patients for three kinds of brain tumor: meningioma (708 slices), glioma (1426 slices), and pituitary tumor (930 slices) (21).The conversion to grayscale from RGB is performed to enrich the images further.The three brain tumor images used are Glioma, Meningioma, and pituitary tumor.As proposed by Demirhan et al. ( 22) converting the input images to grayscale gives a simplified representation is obtained that effectively captures the overall brightness information while eliminating the complexities associated with color.The parameters include boosting the signal-to-noise ratio, making MR images look better, removing the background of unwanted sections, smoothing the inner parts and keeping the essential edges intact.The proposed work is depicted in Figure 1.

Feature extraction
To extract significant features from brain MRI images, six types of feature extraction techniques have been implemented in this study, as listed in Table 1.Using these techniques, important aspects of the images could be identified and analyzed.

GLCM feature extraction
Texture analysis facilitates the differentiation between healthy and unhealthy tissues for visual perception and ML algorithm.In addition, it reveals differences between malignant tumors and normal tissues that might not be observable by the naked eye.By selecting efficient statistical features for early diagnosis, the accuracy can be improved.Second-order statistical texture features can be extracted using GLCM.Creating a GLCM matrix and then deriving statistical metrics from this matrix measures the frequency with which pairs of pixels with specific values and a specific spatial relationship appear in an image.GLCM, or Gray-level spatial dependence matrix (GLSDM), has been used in this research used to extract the statistical features.GLCM was first proposed by Haralick et al. (23) for describing the geographical relationship between pixels with different levels of Gray-level.
To analyze an image's texture statistically, the GLCM counts how often pairs of pixels with the same value and the same relative position appear in the image.By determining the frequency with which pairs of pixels with a given weight and in a given spatial relationship occur in an image, the GLCM functions can characterize an image's texture through the extraction of statistical measurements.The Gray-Level Co-occurrence Matrix (GLCM) is a two-dimensional histogram where each pair of 'p' and 'q' represents the frequency with which the events 'p' and 'q' occur.As a function of distance S = 1, angle (0 degrees horizontal, 45 degrees positive diagonal, 90 degrees vertical, and 135 degrees negative diagonal), and gray scales 'p' and 'q', it determines the frequency with which a pixel of intensity 'p' occurs in proximity to a pixel of intensity 'q' at a given distance 'S' and orientation.
After computing GLCM, five different statistical features are extracted from GLCM.These extracted features include, Contrast: Determines the local variances of the grey-level cooccurrence matrix.
Homogeneity: Determines the proximity of the GLCM element distribution to the GLCM diagonal.
Dissimilarity: Quantifies the range of grayscale intensity.
Energy: It will calculate the pixel's uniformity.
Correlation: Calculates the average degree to which each pixel in the image is connected with its neighbors.
The formulas used to calculate the above characteristic features are shown in Table 2.

LBP features
The Local Binary Patterns (LBP) technique is a widely employed texture descriptor in image processing and computer vision.Local texture representation is a straightforward yet efficient method used to depict the texture characteristics of an image.This Overall proposed model architecture.
technique has been extensively applied in various domains, such as object recognition, face detection, and image segmentation.The step wise LBP Feature Extraction is, each pixel is compared with its neighboring pixels.Consider the Gray value of the center pixel as g c , and the gray value of the neighboring pixel as g p .Comparison is carried out with using Equation 1.
A circle of radius R is centered around the center pixel and the function S is applied to P evenly spaced pixels.

Interaction features
A novel approach implemented in this research is the extraction of image features for tumor classification, wherein a strategy for generating interaction features is employed with the aim of potentially improving the performance of the model.This approach entails the integration of two sets of important attributes: the GLCM attributes and LBP attributes, both of which effectively capture fundamental properties of the images being analyzed.
The GLCM features provides insights into the spatial interdependence of pixel intensities within an image, effectively capturing and representing texture details.In contrast, the features of LBP provide a quantification of the local spatial patterns of pixel luminance, thereby providing supplementary information regarding texture.
The computation of interaction features involves the outer product of the feature vectors obtained from GLCM and LBP.In mathematical terms, the outer product of the GLCM features vector (g) and the LBP features vector (l) yields a matrix (M).Each element M i,j of this matrix represents the product of the i th GLCM feature and the j th LBP feature.The aforementioned statement aptly describes the correlation between the respective GLCM and LBP features.
Subsequently, the interaction matrix is transformed into a onedimensional vector, thereby generating a novel set of features that capture the interplay between the initial feature sets.The inclusion of this extended feature set, in combination with the existing GLCM and LBP features, offers a more exhaustive and refined depiction of the image.Consequently, it has the potential to improve the machine learning model's capacity to differentiate between various tumor classifications.
The steps involved in interaction features are, Defining GLCM and LBP features vector: GLCM feature vector is defined as ɡ = ɡ 1 , ɡ 2 , ɡ 3 ,…ɡ m and the LBP feature vectors as l = l 1 ,1 2 ,l 3 ,… l n in which 'm' is the number of GLCM features and 'n' is number of LBP features.
Calculating the outer product: The matrix 'M' is obtained by computing the outer product of the two given vectors.The matrix is provided as follows: Each element of this matrix M ij = ɡ i l j is a new feature created which captures the interaction between i th GLCM feature as well as with j th LBP feature.
Combining interaction features and original features: The original GLCM and LBP features are combined with the interaction features to create the final feature vector for each image.If the GLCM feature vector 'g' has a size of m and the LBP feature vector 'l' has a size of n, then the final feature vector will have a size of m + n + mXn.
These procedures describe how interaction features are generated from GLCM and LBP features.This augmented feature set has the potential to provide a more comprehensive and nuanced image representation, thereby enhancing the performance of the image classification model.
Here GLCM calculates at 4 different angles such as (0, 45, 90, and 135 degrees).The GLCM algorithm proceeds by computing five distinct properties, namely 'contrast', 'dissimilarity', 'homogeneity', 'energy', and 'correlation', for each angle.These properties are then used to generate a vector of GLCM features.Assuming the vector representing the GLCM features for an image is [10,20,30,40,50].The aforementioned values indicate that the contrast is 10, the dissimilarity is 20, the homogeneity is 30, the energy is 40, and the correlation is 50.
The LBP algorithm is employed to calculate the LBP values using 8 sampling points positioned evenly along a circle with a radius of 1. Subsequently, the histogram of these LBP patterns is computed.Suppose the Local Binary Patterns (LBP) features for a given image are represented by a histogram consisting of 256 bins.The values within this histogram range from 0.01 to 0.01, with each bin containing a distinct value.
The interaction features are generated through the computation of the outer product between the feature vectors of the Grey Level Co-occurrence Matrix (GLCM) and the Local Binary Patterns (LBP).
For the purpose of explanation, consider a simplified scenario where only first five bins of LBP features.These bins are represented by the values [0.01, 0.02, 0.03, 0.04, 0.05].
The 5x5 matrix is obtained by computing the outer product of the GLCM features [10,20,30,40, 50] and the first 5 LBP features [0.01, 0.02, 0.03, 0.04, 0.05].The value of each element in this matrix is obtained by multiplying a feature from the Grey Level Cooccurrence Matrix (GLCM) with a feature from the Local Binary Patterns (LBP) feature.
The 5x5 matrix is obtained by taking the outer product of the GLCM features [10,20,30,40, 50] and the first 5 LBP features [0.01, 0.02, 0.03, 0.04, 0.05].The value of each element in this matrix is obtained by multiplying a feature derived from GLCM with a feature derived from the LBP algorithm as represented in Equation 3.
The interaction features offer way to capture the potentially significant connections among various elements of an image's texture.The precise dynamics of these interactions are dependent upon the data and the attributes of the images under examination.In the context of tumor images, these interactions may potentially expose complicated patterns that are pivotal in distinguishing between various tumor types.

Aggregated features
The aggregated features represent the additional statistical features of GLCM features.The GLCM features can be consolidated into more straightforward statistical measures.The mean, median, and total of the GLCM characteristics are these statistical measurements.The steps involved in aggregated features calculation is shown in Algorithm 1.The features extracted from aggregated features are as follows, Sum: The total number of GLCM characteristics is determined.This gives us a single number that can be taken as a measure of the "amount" of GLCM features in the image.If the GLCM characteristics are represented by the vector g = [ɡ1, ɡ2,…ɡn] then ɡ_m = ɡ1 + ɡ2 + … + ɡn gives the total.Mean: The average GLCM characteristics are determined.This gives us a single number that stands in for the "typical" value of the GLCM feature in the image.
Median: GLCM features median are computed.When the GLCM features are ordered numerically, this yields a single value that characterizes the "middle" value.The median is the midpoint if 'n' is an odd number.If 'n' is divisible by 2, then the median is the midpoint between those two values.
Following these computations, the aggregated GLCM features take the form of a vector with three elements, which are denoted by the notations [ɡ_sum, ɡ_mean, and ɡ_median].
The aggregated features (sum, mean, and median) provide a summary of the GLCM features, capturing various aspects of their level and distribution.In addition to the GLCM features, LBP features, interaction features, statistical features, and non-linear features, these characteristics are added to the final feature vector for each image.
Pattern recognition, machine learning, and image classification employ aggregated features such as summation, average, and median.The raw features are condensed, effectively representing the central tendency and overall pattern of the data, thereby offering a simplified yet meaningful perspective of the dataset.The aforementioned properties exhibit a lower degree of variation compared to individual data points, thereby enhancing the robustness of models against the presence of noisy or outlier data.The utilization of these techniques results in a reduction of data dimensionality, thereby enhancing the efficiency of algorithms.The inclusion of aggregated features in a model has been observed to enhance its predictive performance by uncovering latent data patterns.The code provided utilizes the GLCM attributes to generate aggregated features that summarize the textural characteristics of the image.This has the potential to enhance the tumor classification model.

Statistical GLCM features
A set of statistical measures derived from GLCM of an image constitutes the statistical GLCM features.The GLCM matrix depicts the spatial relationship among image pixel pairs.Each element of the GLCM indicates the probability that two pixels with a particular grey level will occur at a particular distance apart.GLCM statistical traits are used to describe an image's texture (24).Texture is how the pixels in an image are placed in space.It can be used to tell the difference between different kinds of images, like images of nature, medical images, and images of factories, etc.
The following features are derived under statistical GLCM features, Variance: The variance quantifies the degree to which GLCM features deviate from their mean value.When the variance of the GLCM features is high, the values they take on span a wide range, whereas when it's low, the features tend to cluster tightly around the mean.Equation 2 represent the variance formulation.Provided that GLCM feature vector ɡ = [ɡ 1 , ɡ 2 …ɡ m ].Where 'm' is the number of GLCM features.
Where, average of GLCM features is mean.Skewness: The concept of skewness pertains to the degree of asymmetry exhibited by the distribution of GLCM features in relation to their mean value.Skewness is computed as Equation 5, Kurtosis: Kurtosis assesses the tailedness of GLCM feature distributions.High kurtosis shows thick tails and a strong peak, indicating many outliers.Low kurtosis shows light tails and a flat peak, indicating no outliers.Mathematically Kurtosis is expressed in Equation 6.
From Equations 3, 4 'm' is the number of GLCM features and ɡ i refers to the i th element of GLCM feature vector.
These statistical GLCM characteristics can shed light on how the GLCM features are typically distributed.Statistical GLCM features exhibit information on the texture's variability, asymmetry, and outliers, while GLCM features themselves capture the texture details within the image.
The incorporation of statistical GLCM features, such as variance, skewness, and kurtosis, enhances the predictive model by encompassing supplementary distributional information pertaining to the GLCM features.The statistical measures employed in this study shed light on subtle texture variations that may not be easily distinguishable solely from the raw GLCM features, as they effectively capture the spread, asymmetry, and tailedness of these features.The model's exceptional performance metrics, such as its nearly perfect accuracy, precision, recall, and F1 score, are likely enhanced by the incorporation of statistical GLCM features, although other factors may also contribute to these outcomes.The inclusion of these features enhances the diversity of the 1547-feature set, thereby augmenting the model's capacity to accurately classify the tumor images.Statistical GLCM features play a crucial role in enhancing the predictive performance of the model by providing detailed information regarding the texture characteristics of the image.

Non-linear features
Another set of novel features of this research work, in terms of non-linear features, is computed from the GLCM feature vectors by applying a logarithmic transformation.This process generates nonlinear features from the GLCM feature vectors.These GLCM feature vectors, obtained from grayscale image analysis, contain numerical values representing various statistical measures.The application of the logarithmic transformation on these GLCM feature vectors gives rise to the non-linear features.These nonlinear features capture intricate patterns and relationships in the data, which may enhance the performance of machine learning models.
The procedure for obtaining non-linear features from the features derived from the Grey Level Co-occurrence Matrix (GLCM) is as follows, GLCM feature computation: The initial stage entails the computation of the Gray-Level Co-occurrence Matrix (GLCM) features, which serve to extract texture information from the image.The aforementioned features consist of contrast, dissimilarity, homogeneity, energy, and correlation.The computations pertain to the manipulation of the Grey Level Co-Occurrence Matrix (GLCM), a matrix that denotes the occurrence frequency of various combinations of pixel intensities within the image.
Applying Non-linear Transformation: Once the GLCM features have been computed, a non-linear transformation is applied to these features by utilizing the natural logarithm function.The transformation is represented by, From Equation 7where, 'm' as the scalar value or the array of the input for which the natural log is to be determined and the input array is being incremented by '1' which is referred as 'm+1' in Equation 3. Later, the modified array log function is '(m+1)' which is being computed for each element present in GLCM vector.The resultant of this log function is the non-linear features.
The computation of the non-linear features corresponding to each GLCM feature is performed in the following manner.
In the given transformation, it is crucial to add one to the original feature value, denoted as C, before applying the natural logarithm function, represented as C' = log(1+C).The need for this adjustment arises in cases where the correlation coefficient 'C' is equal to zero, as the natural logarithm of zero is undefined.This lack of definition can result in computational challenges.The inclusion of a constant term enables the logarithmic transformation to be applicable to all conceivable values of 'C', encompassing the value of zero.This step is crucial in ensuring the integrity and precision of the feature engineering process.The identical process is employed for the remaining GLCM features, ensuring the integrity of all modified features.
The primary objective of employing the logarithmic transformation is to effectively capture and accentuate non-linear relationships and variations present within the GLCM features.Through the utilization of the logarithm function, the feature values undergo a transformation, resulting in their representation on a logarithmic scale.The utilization of this technique can facilitate the identification of patterns, intensify the differentiation between elements, and enhance the accuracy of the representation of the characteristics of the GLCM.
Non-linear transformations have the potential to effectively capture complex structures within the data, which may not be easy to identify through the original features.Consequently, the utilization of such transformations has the capacity to enhance the performance of machine learning models.

Concatenated features
Finally, the various features including GLCM, LBP, interaction, aggregated, statistical, and non-linear features are consolidated into a unified feature vector for every image.The process involves arranging all the features consecutively to create a lengthy vector.The concatenated feature vector serves as a representation of the image within the feature space, enabling its utilization in subsequent analysis or machine learning endeavours.

Results & discussion
The brain tumor classification model was implemented in Python.Statistical and non-linear feature extraction were used in conjunction with GLCM, LBP, Interaction, and aggregation to create the model.Finally, we classified brain tumors using a support vector machine.
Proposed Composite Feature Extraction Model Performance In this work, a set of performance metrics have been computed to evaluate the effectiveness of the proposed composite Feature Extraction model, as illustrated in Table 3.
Table 4 presents a comprehensive evaluation of performance metrics for different models, including the Composite Feature Extraction model proposed in this study.The assessed metrics include Precision, Recall (or sensitivity), F1-score, Specificity, and Accuracy.
The proposed Composite Feature Extraction model demonstrates exceptional performance in brain tumor classification, achieving approximately 99.83% across all key metrics: Precision, Recall, F1-score, Specificity, and Accuracy.The obtained result indicates the model's high reliability in accurately identifying tumors and healthy cases, effectively reducing false outcomes.The balanced F1-score shows the model's consistent performance in precision and recall.Overall, the model's robustness and superior performance signify its effectiveness in brain tumor classification, surpassing existing models.
The presented Table 5 provides a comparative analysis of the Dice Similarity Coefficient (DSC) values of different models utilized to classify three distinct types of brain tumors, namely glioma, meningioma, and pituitary tumors.The Dice similarity coefficient (DSC) is an essential metric to quantify the degree of similarity between the predicted and actual tumor regions observed in brain images.A greater DSC (Dice similarity coefficient) value indicates a higher level of accuracy in the model, particularly in accurately classifying and distinguishing glioma, meningioma, and pituitary tumors.As proposed, the Composite Feature Extraction model demonstrates exceptional performance with a Dice Similarity Coefficient (DSC) value of 99.6.This value signifies the model's superior accuracy in effectively classifying the three distinct types of brain tumors, surpassing the performance of existing models.
Table 6 illustrates the Composite Feature Extraction model's enhanced efficacy in classifying tumors.The model demonstrates superior performance compared to existing models, as evidenced by its remarkably low False Positive Rate (FPR) of 0.00625 and a False Negative Rate (FNR) of 0.0.These results highlight the model's exceptional accuracy and reliability in effectively reducing false alarms and missed detections.
Table 7 presents the categorization of Meningioma, Glioma, and Pituitary Tumours utilizing various machine-learning techniques.The combined studies used a total of 3064 samples.

Conclusion
Based on the obtained results, the classification model demonstrates exceptional performance, with an accuracy, precision, recall, and F1 score that are all close to 1, indicating a high success rate in classifying across all three classes of brain tumors (Glioma, Meningioma, Pitutary).The model also demonstrates a 100% True Positive Rate (no false negatives) and

Author
Technique used Accuracy (%) This work presents a novel methodology that offers an improved representation of tumor image textures by combining LBP and GLCM features.The addition of interaction features, which are created by taking the outer product of the GLCM and LBP vectors, greatly improves the derived features' ability to discriminate.This unique feature distinguishes the strategy from other approaches.Furthermore, an even more thorough representation of critical tumor image characteristics is obtained through the integration of aggregated, statistical, and non-linear information.Combining the approach with a linear support vector machine classifier yields 99.84% of accuracy rate.We opted to use the Figshare dataset so as to compare our classification accuracies with existing research done with same dataset.Being better than other research outcomes, this method can be applied to real time data.
As the most time-consuming phase, future research could concentrate on streamlining the extraction of features.In addition, future research may investigate the possibility of utilizing the extracted detailed features with less computational time.Another objective of future research is to analyze and classify high-grade brain tumors comprehensively utilizing high-grade brain tumors dataset to improve an understanding of complex tumors.

1 : 4 : 8 :
Input: GLCM Features 2: Output: Aggregated Features 3: Initialize sum as 0 for each GLCM feature in the input do 5: Add the GLCM feature to the sum 6: end for 7: Compute mean as sum divided by the total number of GLCM features Sort the GLCM features in ascending order 9: Compute the median based on the sorted list of GLCM features 1 0 : O u t p u t t h e s u m , m e a n , a n d m e d i a n a s t h e aggregated features Algorithm 1. Calculation of aggregated features.
during the extraction process.However, the training process of the Support Vector Machine (SVM) classifier using these features was completed in a mere 0.16 seconds.The utilization of resources in this study was primarily focused on feature extraction rather than model training.With a strikingly low ratio of 0.00024389, the outstanding efficiency of computation is evident in this scenario, clearly demonstrating the high priority of computing resources.The model attained a perfect average accuracy score of 1.0 on the training data, regardless of the varying sizes of the training sets.Nevertheless, notwithstanding the flawless score, the model exhibited commendable performance on the validation data.The predictive capabilities of the model exhibit a high level of robustness, as indicated by a mean accuracy score that falls within the range of 0.993 to 0.998.Furthermore, it is worth noting that the model consistently demonstrated strong performance across a diverse set of data folds, as indicated by the narrow range of standard deviations, which varied between 0.013 and 0.035.
(13)ure features based on GLCM were first introduced by Haralick et al.(13)in 1979.In biomedical field, advantage of the textural properties of images aids in image classification.

TABLE 1
Feature etxtraction Techniques used in this research work.

TABLE 3
Performance metrics formula.

TABLE 4
Comparison of performance metrics: existing models with composite feature extraction (proposed model).The bold letter used are highlighting the proposed model used in the manuscript.
The bold letter used are highlighting the proposed model used in the manuscript.

TABLE 5
DSC comparison: composite feature extraction vs. existing models.The bold letter used are highlighting the proposed model used in the manuscript.

TABLE 8
Accuracy comparison: composite feature extraction vs. existing models.
The bold letter used are highlighting the proposed model used in the manuscript.Dheepak et al. 10.3389/fonc.2023.1248452Frontiers in Oncology frontiersin.orga 0% False Negative Rate (no false positives), in addition to a very low False Positive Rate, indicating an excellent balance between sensitivity and specificity.
FIGURE 2Proposed composite feature extraction model ROC plot.