Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol., 15 September 2025

Sec. Cancer Epidemiology and Prevention

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1596718

This article is part of the Research TopicAdvanced Machine Learning Techniques in Cancer Prognosis and ScreeningView all 10 articles

Comparative analysis of machine learning techniques on the BraTS dataset for brain tumor classification

Shuping WangShuping Wang1Min Li*Min Li2*
  • 1Information Statistics Center, Hubei Cancer Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
  • 2School of Computer Science and Technology, Hubei Business College, Wuhan, China

Introduction: Accurate classification of brain tumors from MRI scans is a critical task for improving patient outcomes. Machine learning (ML) and deep learning (DL) methods have shown promise in this domain, but their relative performance remains unclear.

Methods: This study evaluates several ML and DL techniques using the BraTS 2024 dataset. The models assessed include traditional algorithms such as Random Forest and advanced deep learning architectures including Simple CNN, VGG16, VGG19, ResNet50, Inception-ResNetV2, and EfficientNet. Preprocessing strategies were applied to optimize model performance.

Results: The Random Forest classifier achieved the highest accuracy of 87%, outperforming all deep learning models, which achieved accuracy in the range of 47% to 70%. This indicates that traditional ML approaches can sometimes surpass state-of-the-art DL methods in tumor classification tasks.

Discussion: The findings highlight the importance of model selection and parameter tuning in automated brain tumor diagnosis. While deep learning models are generally considered standard for image analysis, Random Forest demonstrated superior performance in this context. This underscores the need for fine-grained consideration of dataset characteristics, computational resources, and diagnostic requirements.

Conclusion: The study shows that carefully selected and optimized ML approaches can improve tumor classification and support more accurate and efficient diagnostic systems for brain tumor patients.

1 Introduction

The combination of deep learning and machine learning techniques has resulted in advancement of brain tumor segmentation and classification. The application of these approaches has been a revolution in the field of medical imaging primarily by automating the traditional manually driven processes and accuracy of the tumor identification and classification. As demonstrated by U-Net and Convolutional Neural Networks (CNNs) type of deep learning models, tumor regions in 2D MRI images can be accurately delineated with high Dice similarity coefficients for tumor regions (1, 2). In addition to this, mathematical and machine learning methods such as thresholding, K-Means clustering, and CNN are employed to solve the segmentation problem as an optimization problem by using extracted features to precisely pinpoint tumor edges (3). The segmentation process has been further refined by such other advanced image processing techniques as noise reduction, image enhancement and wavelet analysis (4). As for classification, CNNs have been used effectively for differentiating various tumor types with high classification (1, 5). Support Vector Machines (SVMs) and Random Forests have also been used for tumor classification (Figure 1) by finding best separating hyperplanes or decision trees grown in the feature space (extracted using other set of features) (3). New and more sophisticated neural network architectures, such as NeuraClassNet (6) and MDCNet (7) have reached 99.67% accuracies via novel optimization techniques and multi-view analyses. However, these advancements do not guarantee solution, mainly due to data and model complexity. By using the traditional image processing and the contemporary computational intelligence methods, robust solutions for diverse imaging modalities can be achieved by addressing these constraints, thereby enhancing clinical applicability (8).

Figure 1
Flowchart depicting MRI dataset processing for tumor analysis. Steps include loading the dataset, preprocessing and extracting slices, computing tumor volume, and creating binary labels. Model architecture options include Simple CNN, VGG16, VGG19, ResNet50, Inception-ResNetV2, EfficientNet, and Random Forest. Models are trained, performance evaluated, and ROC curves and accuracy visualized.

Figure 1. Workflow diagram for tumor detection.

The steps include loading the MRI dataset, preprocessing the data, computing tumor volume, creating binary labels, and splitting the data into training and testing sets. The model architecture is selected from various options, including Simple CNN, VGG16/VGG19, ResNet50, Inception-ResNetV2, Efficient Net, and Random Forest. The models are trained and subsequently evaluated for performance, with the results visualized through ROC curves and accuracy metrics. Convolutional Neural Networks (CNNs) have become essential for brain tumor classification because they perform automatic feature extraction from medical imaging data. Pre-trained CNN models have received validation from multiple research studies for their effectiveness in this area. In brain tumor classification medical imaging applications VGG-16 demonstrated its effectiveness by reaching a high 98% accuracy rate according to studies by Vellanki et al., 2024 (9) and Vetrivelan et al., 2024 (10). The ResNet architecture, especially ResNet50, produced strong results with validation accuracies reaching up to 89.47% according to Muftic et al. and Padmakala & Maheswari both in 2024. The EfficientNetB0 and EfficientNetB3 models shows effectiveness by achieving high accuracy levels of 98.36% and 99.44%, respectively (1113). DenseNet models have superior performance by achieving validation accuracies up to 95% (14). The pretrained models that use transfer learning mechanism has the capability to reuse features from large datasets thereby providing significant benefits for medical imaging applications that has limited dataset availability. Hybrid and ensemble learning approaches are used to further enhance the classification accuracy by combining the advantages of multiple models. As an example, the PDSCNN-RRELM combines a lightweight parallel depthwise separable convolutional neural network (PDSCNN) with a hybrid ridge regression extreme learning machine (RRELM) and achieves a great average precision of 99.35% (15). The custom four residual deep learning architecture (4RDL-DCNN) optimized by particle swarm optimization achieved high effectiveness with the accuracy of 98.60% (16) was another noteworthy model. Additionally, a classification accuracy of 95% (17) was achieved on a ViT and EfficientNet-V2 ensemble combined through a genetic algorithm. Ensemble methods demonstrated consistently better results than individual models, benefitting from the strength of combining different architectures to obtain better results. Improvement of model performance and building trust rely heavily on explain ability and optimization techniques. The accuracies achieved for various datasets (11) with the integration of the Quantum Genetic Algorithm (QGA) with EfficientNetB0 are 98.36 and 98.25. In a similar manner, the Particle Swarm Optimization (PSO) algorithm is introduced to optimize the feature extraction process in the custom 4RDL-DCNN model obtaining an accuracy of 98.60% (16). Moreover, weights in the combined system of Vision Transformers (ViT) and EfficientNet-V2 models are optimized using the Genetic Algorithm (GA) to achieve 95% classification accuracy (17). The optimization approaches also improve the feature selection and model weight adjustments to yield improved classification performance. Although deep learning models dominate the field, conventional machine learning (ML) techniques have shown potential in brain tumor classification. The Gaussian Process Classifier (GPC), combined with principal component analysis (PCA), has exhibited notable improvements in accuracy, precision, recall, and F1-score (18). Support Vector Machines (SVMs) have also been utilized alongside feature selection methods like the Gini index and mutual information, yielding competitive results (18, 19). Despite their inherent advantages, traditional machine learning (ML) models generally underperform compared to deep learning architectures in image-based classification tasks.

In medical applications, explain ability is extremely important in order to achieve clinical trust in model decisions. To clarify its decision making process, the PDSCNN-RRELM model has been incorporated with Shapley Additive Explanations (SHAP) (15). Moreover, Gradient-weighted Class Activation Mapping (Grad-CAM) has been employed to visualize the decision making of deep learning models in order to increase their interpretability (20). These approaches make deep learning models in clinical environments transparent and reliable. Significant improvement of model performance was achieved through the application of transfer learning and advanced data preprocessing techniques. Pre trained models such as VGG16, Resnet50, EfficientNetB3 were fine-tuned on brain tumor datasets to achieve the state of the art results (12, 13). Additionally, data augmentation techniques such as sparse auto encoder based augmentation and contrast limited adaptive histogram equalization (CLAHE) (16, 20) were used to enhance the diversity of the dataset and improve model generalization. Different performance metrics such as accuracy, precision, recall and F1 score have been used to evaluate various models effectiveness. Vellanki et al. made noteworthy findings, namely that VGG-16 has achieved 98% accuracy in brain tumor classification while Gencer & Gencer obtained 98.36% accuracy over traditional methods with EfficientNetB0. The denseNet has shown good generalization to a validation accuracy of 95% (14). In addition, the accuracies of ensemble models using multiple deep learning architectures (21) can reach up to 99.43%. There has been progress, but there are still a lot of obstacles. It was observed that models such as EfficientNetB3 and VGG-19, when trained on a limited amount of data, exhibited signs of overfitting (12, 13). In order to train more complex models, larger datasets are required (12, 13). Furthermore, most of the models have been tested on the standard datasets, which makes it important to evaluate them further in real clinical settings (14, 22). Other issues should be tackled in future studies and new architectures and optimization methods explored to improve model accuracy and dependability.

The BraTS 2024 database (32) contains T1w, T1w contrast enhancement (T1c), T2w, and FLAIR sequence types, each offering distinct information for imaging various aspects of tumor morphology and structure (23). This research concentrated on utilizing a subset of these modalities, specifically contrast-enhanced T1 (T1c), T2w, and T2-FLAIR (or T2w) images for tumor classification. Additionally, the corresponding segmentation masks were employed to extract quantitative measurements (such as tumor size) used to create binary labels. Patients with tumor volumes exceeding the median were classified as having high tumor burden, while those below the median were categorized as having low tumor burden. For decades, deep learning has the potential to solve the complex patterns in medical images. The architectures used in this study include VGG16 (24), VGG19 (25), ResNet50 (26), Inception-ResNetV2 (27), and EfficientNet (28), all of which are known to achieve state-of-the-art performance in various image classification tasks. Even in this case, classical machine learning techniques, e.g., Random Forests, are very competitive (and in combination with PCA robust feature extraction, which is in particular useful when there is a lack of data and/or the features are very discriminative) (29, 30). Despite numerous studies utilizing these methods, no comprehensive evaluation of this nature on the BraTS 2024 dataset has been reported in the existing literature. This gap is filled by evaluating several state of the art deep learning models as well as Random Forest classifier on PCA reduced features. Additionally, the models were compared in terms of accuracy, loss, and confusion matrices, and the results were visualized at both individual and aggregate levels (4042).

This research has three main goals. First, it evaluates how different deep learning models and a traditional Random Forest classifier perform when applied to the BraTS 2024 dataset. Second, it examines how various data preprocessing techniques and labeling approaches affect model effectiveness. Lastly, this investigation provides valuable insights and suggestions for enhancing brain tumor classification methods. The findings are intended to inform future studies and practical applications of automated brain tumor detection and analysis.

2 Materials and methods

This section describes the BraTS 2024 dataset and outlines the preprocessing pipeline implemented to prepare the data for model training. The focus was on leveraging multimodal MRI scans, extracting relevant features, and generating labels for binary classification.

2.1 Dataset

The BraTS 2024 dataset provides a comprehensive collection of multimodal MRI scans of patients with brain tumors https://arxiv.org/abs/2405.18368 (32). The dataset was obtained from a Kaggle repository and organized into patient-specific folders. For example, a typical folder (for example, BraTS-GLI-02632-102) contains files named according to the following pattern:

● BraTS-GLI-02632-102-t1c.nii

● BraTS-GLI-02632-102-t2w.nii

● BraTS-GLI-02632-102-t2f.nii

● BraTS-GLI-02632-102-seg.nii

This hierarchical structure facilitates individual patient-level processing, ensuring that each modality and its corresponding segmentation mask are correctly associated with each other. Harmonizing data for training requires a preprocessing step. For the data preprocessing pipeline, nibabel library was used to load the NIfTI files. Middle slices were extracted from each modality (T1c, T2w, and T2-FLAIR) assuming that they represent the tumor region. Finally, each extracted slice was resized to 128×128 pixels using the resize function from skimage. Transform and normalized to the [0,1] range to reduce variability across scans. A stack of 3 channel image was formed for each patient by stacking the processed slices from the three modalities along the channel dimension. Then, a segmentation mask was used to compute the tumor volume by counting nonzero voxels, which served as a quantitative measure used to develop classification labels. As there is no explicit clinical labels in the dataset, the binary classification labels were computed based on the tumor volume. More specifically, for all patients the median tumor volume was calculated and if a tumor volume exceeded the median, 1 was assigned as its label (indicative of a high tumor burden). Patients with volumes above the median were labelled with a value of 1, indicating a high tumor burden, and patients with volumes below the median were given a label of 0. It also leads to a balanced binary labelling scheme for the classification task. The data were preprocessed and labeled, and the data were partitioned into training and testing subsets by 80/20 ratio. The partition was done at patient level so that there are no patient data in both the training and the testing sets. This approach helps in keeping the independence of the evaluation data and gives a fair estimate of the model performance. The following section details how the BraTS 2024 dataset was prepared for classification. The first step of preprocessing made sure that the imaging data were the same size and intensity, and the second step ensured that the label generation step allowed for a clear definition of classes in terms of the tumor burden. These procedures were used to guide subsequent model training and evaluation as laid out in the next sections. Each patient folder in the dataset contained several NIfTI files corresponding to different imaging modalities, including:

● T1c (Contrast-Enhanced T1): Highlights regions with blood–brain barrier disruption.

● T2w (T2-Weighted): Provides a high signal intensity for fluid regions.

● T2-FLAIR (Fluid-Attenuated Inversion Recovery): Suppresses cerebrospinal fluid signals to better visualize the lesions.

● Segmentation Masks: Contain expert annotations delineating the tumor regions.

The T1c, T2w, and T2-FLAIR modalities were utilized to construct a 3-channel input image for each patient, while the segmentation mask was used to derive quantitative measures, such as tumor volume.

Primarily one needs to load the dataset and perform pre-processing on the data and utilize PCA for dimensional reduction. The information undergoes division into training batches and testing batches. The Random Forest model receives its training through application of the training data. The testing data receives predictions from the model and its performance undergoes evaluation. The visual representation includes the ROC curve.

2.2 Methodology

This section outlines the overall experimental design, including the models evaluated, training procedures, and the evaluation metrics and visualizations used to compare performance (Figure 2). To capture the intricate patterns, present in brain MRI data, several state-of-the-art deep learning architectures were implemented using transfer learning. The foundational baseline model employed was a Simple CNN, representing a custom-designed convolutional neural network. Subsequently, the VGG16 architecture, consisting of 16 layers, was utilized due to its suitability for transfer learning by freezing the convolutional base and appending custom fully connected layers. This was followed by the use of VGG19, which extends the architecture to 19 layers to enable the extraction of more abstract features. Additionally, ResNet50, a 50-layer residual network with shortcut connections, was incorporated to facilitate the training of deeper models. Inception modules and the residual connection-based advanced hybrid architecture, InceptionResNetV2, were introduced to capture multi-scale features. Lastly, EfficientNet was employed, which balances model depth, width, and resolution through compound scaling, representing a modern and efficient network design. In neural models, an increase in the number of layers generally results in a higher number of trainable parameters (weights and biases), allowing the network to model more complex relationships. However, deeper architectures are more prone to overfitting, particularly when the dataset is limited, and thus require additional regularization and tuning strategies to generalize effectively (38, 39). A Random Forest classifier is the classical machine learning approach which, in addition to deep learning models, was used for classification. Then, for this method, image detecting features were extracted through flattening the pixel intensities and followed by the dimensionality reduction method of Principal Component Analysis (PCA). Under conditions where the dataset size was limited, this classical technique acted as a benchmark for the deep learning models.

Figure 2
Flowchart detailing a machine learning process for MRI data. Steps include loading MRI datasets, preprocessing, computing tumor volume, creating binary labels, selecting model architecture from options like Simple CNN and ResNet50, training models, evaluating performance, and visualizing ROC curves and accuracy.

Figure 2. A flowchart presents the complete procedure for data analysis.

In this study on brain tumor classification using the BraTS 2024 dataset, a Random Forest classifier was employed to distinguish between cases of high and low tumor volumes. Prior to classification, each patient’s multi-modal brain MRI comprising t1c, t2w, and t2f sequences is preprocessed and flattened into a high-dimensional feature vector x. To reduce dimensionality and mitigate overfitting, Primary Component Analysis (PCA) is applied to project these feature vectors onto a lower-dimensional space.

This transformation is mathematically expressed in Equation 1 as:

z=wT(xx)(1)

Where x represents the mean of the training data and W is the projection matrix containing the top principal components. The transformed feature vector z thus captures the most significant variance in the data while discarding redundant information.

The Random Forest classifier then operates on these PCA-reduced features. Let (z) denote the prediction made by the kth decision tree in the ensemble for the transformed input z. In a binary classification setting (with class labels 0 for low tumor volume and 1 for high tumor volume), the final prediction y^  is obtained through a majority voting scheme as presented in Equation 2:

y^ = modeT1z, T2z, , TKz(2)

This process can also be formulated in Equation 3 as:

y^=arg maxck=1K1{Tk(z)=c}(3)

Where it1=c/it  is an indicator function that equals 1 if the kth tree predicts class c and 0 otherwise.

Alternatively, if each tree provides a probabilistic estimate pk for class c, the ensemble’s class probability is calculated as the average of these estimates as shown in Equation 4:

p^ cx= 1kk=1Kpk cx(4)

The final class label is then determined by selecting the class with the highest average probability as presented in Equation 5:

y =argmaxc p^ (cx)(5)

2.2.1 Data preprocessing pipeline

In the context of the BraTS2024 dataset, this approach leverages the multi-modal imaging data by first condensing it via PCA and subsequently classifying patients based on tumor volume. The middle slice from each modality (T1c, T2w, and T2-FLAIR) was extracted and resized to 128×128 pixels. The three modalities were stacked to form a three-channel image, with each channel corresponding to a specific modality. All images were normalized to the [0, 1] range to maintain consistency across models.

2.2.2 Training configuration

The dataset was split into training and testing sets using an 80/20 ratio, ensuring no patient appeared in both sets. Data augmentation techniques (rotation, flip, and scaling) were applied to enhance generalization. The Adam optimizer was employed with learning rates (1e-4 to 1e-3), and Categorical Cross-Entropy served as the loss function for binary classification with one-hot encoded labels. Training used a batch size of 8 and initially spanned five epochs, extended as needed for improved performance.

2.2.3 Model architectures and training

For deep learning models, the convolutional bases of pretrained models were frozen during initial experiments, with options for unfreezing and fine-tuning in future trials. For the Random Forest approach, images were flattened, and PCA reduced feature dimensionality (minimum of 50 components). The Random Forest classifier was trained with 100 estimators and default parameters as a classical benchmark.

2.2.4 Performance evaluation

Model performance was assessed through:

● Training/validation accuracy and loss curves across epochs

● Confusion matrices analyzing true/false positives/negatives

● ROC curve analysis with AUC values

● Aggregate plots comparing final accuracies and losses

● Grid layouts of all confusion matrices

This comprehensive evaluation framework enabled robust comparison of both deep learning and classical machine learning approaches on the BraTS 2024 dataset.

3 Results

This section presents the experimental results of the BraTS dataset 2024 used in this study. Deep learning models and classical Random Forest classifier are evaluated for comparison. The performances of both individual and aggregated models are visualized in terms of accuracy, loss and confusion matrices. The experiments are evaluated using the BraTS 2024 dataset. The data in the dataset is preprocessed for each patient by extracting the middle slices of three imaging modalities (T1c, T2w, T2-FLAIR) and normalizing them. To create binary labels from the tumor volume computed based on the segmentation masks, the median volume was chosen as the threshold (Table 1). The following accuracy results were obtained for the test set:

Table 1
www.frontiersin.org

Table 1. Obtained results of accuracy.

The Random Forest model with PCA features had the highest accuracy of 87.5%. More complex CNN models performed with lower accuracy rates. The results suggest that the evaluated deep learning models failed to match the performance of the classical Random Forest classifier by a large margin. On the other side of comparison, different deep models showed more ranges of performance, simple architecture (Simple CNN and VGG based networks) got higher accuracy than complex networks (ResNet50 and EfficientNet) (Table 2, Figure 3).

Table 2
www.frontiersin.org

Table 2. Performance metrics for a Random Forest classifier.

Figure 3
Bar chart titled “Random Forest Performance Metrics” displaying scores for five metrics: Accuracy at 0.88, Precision at 0.90, Sensitivity (Recall) at 0.86, Specificity at 0.89, and F1 Score at 0.88.

Figure 3. Performance metrics on the BraTS 2024 dataset.

Table 2 displays the model’s overall accuracy (0.88) along with more detailed measures: precision (0.90) indicates how many of the positive predictions were correct; sensitivity or recall (0.86) shows the model’s ability to correctly identify actual positive cases; specificity (0.89) reflects how well the model identifies negative cases; and the F1 Score (0.88) provides a balance between precision and recall. Together, these metrics offer a comprehensive overview of the classifier’s performance. The key performance metrics for the Random Forest classifier offer a comprehensive overview of the model’s effectiveness. The overall accuracy is 0.88. More detailed measures include precision at 0.90, indicating the proportion of correct positive predictions; sensitivity (or recall) at 0.86, reflecting the model’s ability to identify actual positive cases; specificity at 0.89, demonstrating how well the model recognizes negative cases; and an F1 Score of 0.88, which balances precision and recall. Together, these metrics clearly and effectively illustrate the classifier’s performance.

The graph displays an accuracy of 0.88, precision of 0.90, sensitivity (recall) of 0.86, specificity of 0.89, and an F1 score of 0.88. These metrics collectively demonstrate that the classifier effectively balances true positive and true negative rates, ensuring robust discrimination between high and low tumor burden cases. The visualization underscores the efficacy of the traditional Random Forest approach, particularly when leveraging PCA-reduced features for brain tumor classification. Each model was trained and validated independently in their respective loss curves and plotted the training and validation accuracies, and revealed themselves to be very good at learning. Accuracy of the Simple CNN increased with each training epoch and corresponding loss decreased. Same trends were observed in VGG16 and VGG19, but with slightly worse final accuracies (Figure 4). In contrast, more complex architectures such as ResNet50 and EfficientNet exhibited poor convergence, as indicated by static accuracy curves and relatively high loss values (Figure 5). At the same time, the better performance of the shallow architectures was complemented by moderate improvement of the InceptionResNetV2 model, which, however, could not achieve performance levels comparable to the straight architectures.

Figure 4
Four line graphs comparing training and validation metrics over epochs for Simple CNN and VGG16 Transfer Learning models. (a) Simple CNN Accuracy shows both training and validation accuracy increasing, with training accuracy higher. (b) Simple CNN Loss depicts decreasing loss, with training loss lower. (c) VGG16 Transfer Learning Accuracy shows an increase in accuracies, with training accuracy higher. (d) VGG16 Transfer Learning Loss shows decreasing loss, with training loss lower.

Figure 4. Performance metrics for the simple CNN and VGG16 transfer learning models over four epochs: (a) Training and validation accuracy (b) Training and validation loss (c) Training and validation accuracy for the VGG16 transfer learning model (d) Training and validation loss for the VGG16 model.

Figure 5
Six graphs compare the performance of ResNet50, Inception-ResNetV2, and VGG19 models. Each model has accuracy and loss charts. Accuracy generally increases over epochs, while loss decreases. Training and validation trends are visualized separately for each model.

Figure 5. Performance metrics for ResNet50 and Inception-ResNetV2 models over four epochs. (a) Training and validation accuracy for the ResNet50 model (b) Training and validation loss for the ResNet50 model (c) Training and validation accuracy for the Inception-ResNetV2 model (d) Training and validation loss for the Inception-ResNetV2 model (e) Training and validation accuracy for the VGG19 transfer learning model (f) Training and validation loss for the VGG19 model.

The accuracy and loss curves provide valuable insights into the training progress and performance of both Simple CNN and VGG16 models. Accuracy curves plot training and validation accuracy against the number of epochs, where higher accuracy indicates better prediction capability. Similarly, loss curves depict the model’s error over time, with separate plots for training and validation loss. A decreasing loss suggests improved performance and more accurate predictions. Comparing the two models, the graphs highlight differences in generalization, with VGG16, a pre-trained model, typically converging faster and achieving higher accuracy than Simple CNN, which is trained from scratch. Observing the intersection points and divergence of curves helps evaluate which model performs better on unseen data, demonstrating the advantage of transfer learning in achieving efficient and effective model training.

Three neural network architectures evaluate their training performance through multiple epochs in the figure. The accuracy plots show training and validation results that might reveal overfitting or under fitting symptoms through their changing patterns. Loss graphs demonstrate the training error reduction pattern of the model throughout its learning process. The large difference between training and validation accuracy can indicate pronounced overfitting when training deeper networks such as ResNet50 and Inception-ResNetV2. The stable performance of VGG19 results from transfer learning because it uses pre-trained models to achieve improved generalization capabilities on unknown data.

The experimental training performance of the EfficientNet model, including trends in accuracy and loss over training epochs, is presented in Figure 6. The illustration in this figure shows how the Efficient Net model performed throughout its training process. During training the accuracy graph demonstrates model learning effectiveness through increasing training accuracy levels and occasionally fluctuating validation accuracy which might lead to overfitting because the model adapts to training data patterns. This graph shows error rate measurements alongside the training process which indicates that effective error reduction occurs when loss values decrease. Each model was used to generate its own class specific confusion matrix. Overall accuracy for the Random Forest classifier was quite high, and the confusion matrix of the classifier revealed strong discrimination between the two classes. A balanced distribution of true positive vs. true negative was observed for the Simple CNN and VGG16 although some misclassifications could be found (Figure 7). On the contrary, models including ResNet50 and EfficientNet had slightly higher confusion rate between the classes, which matched their slightly lower accuracy values for a couple of seconds (Figure 8). The Random Forest classifier had high overall accuracy and had a robust discrimination of the two classes in the confusion matrix of the predicted model. The true positives and true negatives were distributed balanced with some misclassifications for the Simple CNN and VGG16 models. On the other hand, ResNet50 and EfficientNet showed higher level of confusion between classes as they had lower accuracy scores. The confusion matrices for various models provide a detailed view of their classification performance by comparing the true labels with the predicted labels. In the case of the Simple CNN model, the confusion matrix (a) displays the distribution of predictions across different classes, indicating a moderate performance with some misclassifications. The VGG16 model (b) shows improved classification accuracy, as reflected in its confusion matrix, where most predictions align correctly with the actual classes. The VGG19 model (c) further enhances this trend, demonstrating strong classification performance with very few misclassified instances, suggesting better feature extraction and generalization. Lastly, the ResNet50 model (d) stands out with its confusion matrix revealing high classification accuracy and minimal errors, underlining the effectiveness of residual connections in preserving learned features and improving model depth without performance degradation.

Figure 6
(a) Line chart showing EfficientNet accuracy over epochs. Train accuracy peaks at epoch 2 before declining. Validation accuracy peaks at epoch 3 then decreases. (b) Line chart of loss values. Train loss decreases steeply at first, then fluctuates. Validation loss increases slightly then decreases.

Figure 6. Performance metrics for the EfficientNet model over four epochs (a) Training and validation accuracy (b) Training and validation loss.

Figure 7
Four confusion matrices compare model predictions. (a) Simple CNN: True positives 10, false positives 9, false negatives 3, true negatives 18. (b) VGG16: True positives 19, false positives 0, false negatives 13, true negatives 8. (c) VGG19: True positives 19, false positives 0, false negatives 15, true negatives 6. (d) ResNet50: True positives 19, false positives 0, false negatives 21, true negatives 0. Each matrix has predicted labels on the x-axis and true labels on the y-axis.

Figure 7. Confusion matrices for various models illustrating their classification performance: (a) For Simple CNN model (b) For the VGG16 model.(c) For the VGG19 model (d) For the ResNet50 model.

Figure 8
Three confusion matrices comparing classification models. The first, for Inception-ResNetV2, shows 5 true negatives, 2 false positives, 14 false negatives, and 19 true positives. The second, for EfficientNet, shows 19 true negatives, 23 false positives, and no true positives or negatives. The third, for Random Forest, shows 17 true negatives, 3 false positives, 2 false negatives, and 18 true positives. Each matrix includes a color gradient representing values.

Figure 8. Confusion matrices of Inception-ResNet V2, Efficient Net and Random Forest classifiers.

The confusion matrix displays quantitative data about correct and incorrect predictions between categories which enables detailed strength and weakness assessments of each predictive model. Superior model accuracy correlates with the number of correctly identified instances that appear along the diagonal of the confusion matrix yet the off-diagonal elements show misclassified classes. The accuracy levels alongside generalization ability of Inception-ResNet V2 and Efficient Net surpass Random Forest classifier because these deep learning models use modern machine learning technology. The developed confusion matrices reveal classification capabilities of each model to help choose the optimal method for the dataset.

3.1 ROC curve analysis

Softmax probabilities were plotted as ROC curves (Figure 9) for each of deep learning models. Additional evidence of performance disparities between the models was obtained through the area under the curve (AUC) values (Figure 10).

Figure 9
Six ROC curve plots compare different models. (a) Simple CNN with AUC of 0.84. (b) VGG16 with AUC of 0.83. (c) VGG19 with AUC of 0.84. (d) ResNet50 with AUC of 0.55. (e) Inception-ResNetV2 with AUC of 0.68. (f) EfficientNet with AUC of 0.61. All plots show true positive rate vs. false positive rate with a diagonal reference line.

Figure 9. Receiver Operating Characteristic (ROC) curves for various models, illustrating their classification performance. (a) ROC curve for the Simple CNN model (b) ROC curve for the VGG16 model (c) ROC curve for the VGG19 model (d) ROC curve for the ResNet50 model. (e) ROC curve for the Inception-ResNetV2 model (f) ROC curve for the EfficientNet model.

Figure 10
ROC curve for a Random Forest model showing the true positive rate against the false positive rate. The curve has an AUC of 0.90, indicating strong model performance. The diagonal line represents a baseline without predictive power.

Figure 10. ROC curve for Random Forest classifier.

The ROC curve for the Random Forest classifier illustrates its ability to distinguish between different classes by plotting the True Positive Rate (sensitivity) against the False Positive Rate at various threshold values. A well-performing model has a curve that leans toward the top-left corner, indicating a high True Positive Rate with a low False Positive Rate. The Area Under the Curve (AUC) serves as a key performance metric, where a higher AUC value signifies better classification capability. Random Forest, as an ensemble learning method, typically achieves a strong AUC by leveraging multiple decision trees to enhance predictive accuracy and reduce overfitting. The ROC curve provides insights into the classifier’s effectiveness and helps in comparing its performance against deep learning models.

3.2 Combined visualization

To enable a holistic comparison across all models, the final performance metrics were aggregated (Figure 11) and visualizations (Figure 12) as follows:

Figure 11
ROC curve comparison chart for various models, displaying True Positive Rate against False Positive Rate. Curve legends with AUC values: Simple CNN (0.84), VGG16 (0.83), VGG19 (0.84), ResNet50 (0.55), Inception-ResNetV2 (0.68), EfficientNet (0.61), Random Forest (0.90). Dashed diagonal indicates random classification.

Figure 11. Combined ROC curve.

Figure 12
Bar chart comparing the accuracy of various models: Simple CNN, VGG16, VGG19, ResNet50, Inception-ResNetV2, EfficientNet, and Random Forest. Random Forest shows the highest accuracy, while ResNet50 has the lowest.

Figure 12. Accuracy comparison of various deep learning models.

Comparison for various classification models, including Simple CNN, VGG16, VGG19, RESNET50, Inception-ResNet V2, efficient Net, and Random Forest is presented in Figure 11. The Area Under the Curve (AUC) values are indicated for each model, showing the model’s ability to distinguish between classes. The dotted diagonal line represents the performance of a random classifier, with an AUC of 0.5.

It is clear from Figure 12 that Random Forest outperforms other models with an accuracy of 87%. Model Accuracy Comparison for Simple CNN, VGG16, VGG19, RESNET50, Inception-ResNet V2, efficient Net, and Random Forest. This bar chart illustrates the accuracy of each model, highlighting Random Forest as the highest-performing model. A bar chart was used to compare the final validation accuracies of all models whereas another bar chart shows the final validation losses of the deep learning models. Overall, these aggregated plots show that the Random Forest classifier achieved the highest accuracy on the whole. They also worked moderately well on deep learning architectures in terms of accuracy and loss metrics and far worse than Simple CNN, VGG16 and VGG19, ResNet50, Inception-ResNetV2 and EfficientNet. In addition to the individual plots, all confusion matrices were arranged into a single grid to facilitate direct visual comparison of class predictions across different models. This consolidated view illustrates that the Random Forest classifier outperforms the deep learning models in terms of class separation and classification accuracy. However, under these preprocessing and training conditions, it was shown that the Random Forest classifier with PCA reduced features outperformed the evaluated deep learning models. This indicates that feature extraction through PCA and the robustness of the classical machine learning methods can be incredibly effective for this task in the presence of a small dataset or when a deep model is not well fine-tuned. Detailed visualizations here help extend and understand each model’s behavior before hyperparameter tuning, fine tuning of pre trained networks and data augmentation strategies to achieve better performance if any of these potential deep learning model performance improvement techniques.

4 Discussion

The integration of deep learning with other machine learning techniques has gained significant attention in brain tumor classification, segmentation, and detection, particularly using MRI datasets, as highlighted in recent studies (31, 32). These hybrid approaches have demonstrated notable potential in achieving both clinical accuracy and computational efficiency. For example, frameworks incorporating the Artificial Intelligence of Things (AIoT) have validated the feasibility of embedding AI models into smart healthcare systems to enable more effective real-time diagnostic support (33). Similarly, traditional machine learning models, when combined with optimized feature extraction strategies, have proven effective in the classification and detection of tumors from MRI scans (34). Comparative and benchmarking studies further emphasize the strong predictive capabilities of classical machine learning algorithms, especially in scenarios involving limited data or high-dimensional input features (35). Moreover, empirical evaluations underscore the continued importance of performance tuning and algorithm selection in refining outcomes within biomedical imaging applications (3642). These findings support the conclusions of the current study and justify the inclusion of both conventional and deep learning models in the experimental framework.

The experimental results revealed significant performance differences between classical and deep learning approaches. Notably, the feature set derived using Principal Component Analysis (PCA) achieved an accuracy of 87.5% with the Random Forest classifier—surpassing all tested deep learning models. In contrast, the Simple CNN reached only 70% accuracy, while other architectures such as VGG16, VGG19, ResNet50, Inception-ResNetV2, and EfficientNet performed even lower, with accuracies ranging from 47.5% to 67.5%. These findings underscore the robustness of PCA-based features in conjunction with classical machine learning algorithms within the current experimental setup. Although deep learning models inherently offer greater complexity and theoretical potential, their performance in this study was hindered by factors such as limited training data, overfitting, and insufficient benefit from transfer learning using pre-trained networks. The processing pipeline played a crucial role in shaping these results. Specifically, it involved extracting the middle slice from each MRI modality, normalizing the images, and employing segmentation masks to compute tumor volumes. Binary labels were then assigned using the median tumor volume as a threshold. While these preprocessing and labeling strategies provided a standardized framework for model comparison, they may also have introduced certain limitations. For instance, relying solely on the middle slice may fail to capture the full spatial heterogeneity of the tumor. Likewise, deriving binary labels based solely on tumor volume oversimplifies the complex clinical distinctions between high-grade and low-grade tumors, which are not strictly dichotomous. Refining these preprocessing and labeling methods could potentially enhance the performance of deep learning models in future studies. The proposed method demonstrates competitive performance when compared with recent approaches in brain tumor classification. As shown in Table 3, our Random Forest model with PCA-based feature reduction achieved superior classification accuracy (87.5%) compared to existing methods: the Probabilistic Neural Network (83.33%) (43) and Vision Transformer (81.5%) (44). This 4.17–6.0% improvement demonstrates the efficacy of our hybrid approach, where PCA optimally preserves discriminative features while Random Forest’s ensemble learning mitigates overfitting—particularly advantageous for limited medical imaging datasets.

Table 3
www.frontiersin.org

Table 3. Comparison with state of the art.

4.1 Potential causes for underperformance of certain deep learning models

The underperformance of certain deep learning architectures—such as ResNet50 and EfficientNet—compared to simpler models like the Simple CNN appears to be driven by several underlying factors. A primary issue is that complex models typically require large and diverse datasets to effectively learn robust feature representations. However, the dataset used in this study was relatively small and sourced from real-world clinical data, which may have constrained the learning capacity of these deeper networks. Additionally, limitations in fine-tuning further contributed to this performance gap. The pre-trained networks were primarily employed as fixed feature extractors, restricting their ability to adapt to the specific characteristics of brain tumor MRI data. To enhance domain-specific learning, selectively unfreezing and fine-tuning deeper layers could enable these models to capture more relevant features. Overfitting was also likely exacerbated by a lack of extensive data augmentation and the application of strong regularization techniques. Implementing a broader range of augmentation strategies such as rotation, flipping, and intensity variation could improve generalization and help mitigate the risk of overfitting in future implementations.

4.2 Limitations and considerations

The current study has several limitations. First, although the dataset comprises real BraTS 2024 images, the effective sample size, particularly after splitting into training and testing sets remains relatively small, potentially preventing more complex deep learning models from achieving optimal performance. Next, the choice to extract only the middle slice from each modality simplifies the data representation but may forfeit valuable spatial information that could enhance the classification. Furthermore, employing tumor volume as the basis for binary labels offers a straightforward mechanism for class differentiation but may not encompass all clinically relevant factors. Future studies should consider alternative or supplementary labelling strategies that integrate additional clinical data. The reliance on fixed, pre-trained models may have limited their potential in the medical imaging context. A more nuanced fine-tuning process tailored specifically to the BraTS dataset could enable better adaptation and improved performance for these architectures. Although the Random Forest classifier demonstrated superior performance compared to the deep learning models in the current analysis, the findings highlight several opportunities for enhancing the effectiveness of deep learning approaches. Addressing these issues through enhanced data preprocessing, fine-tuning strategies, and robust augmentation may bridge the performance gap and lead to more clinically useful models in the future.

5 Conclusion

This study evaluated various machine learning techniques for classifying intra-brain tumors using the BraTS 2024 dataset. The Random Forest model combined with PCA for feature reduction achieved the highest accuracy (87.5%), showing that classical methods with proper preprocessing can be very effective. Some deep learning models, such as ResNet50 and EfficientNet, performed worse, suggesting that these require careful tuning and training strategies to reach their potential. Future work should focus on combining classical and deep learning approaches, including ensemble methods, to improve accuracy and robustness.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author. Requests to access these datasets should be directed to ML, bG05NzkzMDgzNjdAMTYzLmNvbQ==.

Author contributions

SW: Data curation, Formal Analysis, Investigation, Methodology, Resources, Writing – original draft. ML: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Software, Visualization, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Alagarsamy S, Durgadevi V, Shahina A, Anish L, Kameshwar A, and Madeshwaran A. (2024). Brain Tumor Segmentation and Classification using Deep neural networks, in: 2020 4th International Conference on Electronics, Communication and Aerospace Technology (ICECA), . pp. 1124–9. doi: 10.1109/iceca63461.2024.10801098

Crossref Full Text | Google Scholar

2. B. R, Kumar TRD, Ravi VR, V. R. K, V. R. K, and G. R. P. (2024). Segmentation And Classification Of Brain Tumor Using CNN Algorithm, in: 2024 7th International Conference on Circuit Power and Computing Technologies (ICCPCT), Kollam, India. pp. 813–9. doi: 10.1109/ICCPCT61902.2024.10672932

Crossref Full Text | Google Scholar

3. Amin J, Sharif M, Haldorai A, Yasmin M, and Nayak RS. Brain tumor detection and classification using machine learning: a comprehensive survey. Complex Intelligent Syst. (2021) 8(4):3161–83. doi: 10.1007/s40747-021-00563-y

Crossref Full Text | Google Scholar

4. Alam MS, Rahman MM, Hossain MA, Islam MK, Ahmed KM, Ahmed KT, et al. Automatic human brain tumor detection in MRI image using Template-Based K means and improved fuzzy C means clustering algorithm. Big Data Cogn Computing. (2019) 3(2):27. doi: 10.3390/bdcc3020027

Crossref Full Text | Google Scholar

5. Mostafa H, Haddad N, Mohamed H, and Taha ZAEH. (2024). Brain MRI Classification and Segmentation of Glioma, Pituitary and Meningioma Tumors Using Deep Learning Approaches, in: 2024 Intelligent Methods, Systems, and Applications (IMSA), Giza, Egypt. pp. 482–8. doi: 10.1109/IMSA61967.2024.10652735

Crossref Full Text | Google Scholar

6. Jayachandran A and Anisha N. Multi-class brain tumor classification system in MRI images using cascades neural network. Comput Intell. (2024) 40(4). doi: 10.1111/coin.12687

Crossref Full Text | Google Scholar

7. ThamilSelvi C, Vinoth Kumar S, Asaad RR, Palanisamy P, and Rajappan LK. An integrative framework for brain tumor segmentation and classification using neuraclassnet. Intelligent Data Anal. (2024) 29(2):435–58. doi: 10.3233/IDA-240108

Crossref Full Text | Google Scholar

8. Batool A and Byun Y. Brain tumor detection with integrating traditional and computational intelligence approaches across diverse imaging modalities - Challenges and future directions. Comput Biol Med. (2024) 175:108412. doi: 10.1016/j.compbiomed.2024.108412

PubMed Abstract | Crossref Full Text | Google Scholar

9. Vellanki Y, Popuri S, Maram LY, and Kamepalli S. (2024). Enhanced Brain Tumor Detection with Deep Learning: A Comparative Study of CNN, VGG16 and ResNet Models, in: 2024 First International Conference on Innovations in Communications, Electrical and Computer Engineering (ICICEC), Davangere, India, Vol. 2024. pp. 1–8. doi: 10.1109/ICICEC62498.2024.10808608

Crossref Full Text | Google Scholar

10. Vetrivelan P, Sanjay K, Shreedhar GD, S KV, and Nithyan R. (2024). Brain Tumor Detection and Classification Using Deep Learning, in: 2024 International Conference on Smart Systems for Electrical, Electronics, Communication and Computer Engineering (ICSSEECC), Coimbatore, India, Vol. 2024. pp. 1–6. doi: 10.1109/10.1109/ICSSEECC61126.2024.10649538

Crossref Full Text | Google Scholar

11. Gencer K and Gencer G. Hybrid deep learning approach for brain tumor classification using EfficientNetB0 and novel quantum genetic algorithm. PeerJ Comput Sci. (2025) 11:e2556. doi: 10.7717/peerj-cs.2556

PubMed Abstract | Crossref Full Text | Google Scholar

12. Muftic F, Kadunic M, Musinbegovic A, Almisreb AA, and Ja’afar H. Deep learning for magnetic resonance imaging brain tumor detection: Evaluating ResNet, EfficientNet, and VGG-19. Int J Power Electron Drive Syst. (2024) 14(6):6360–72. doi: 10.11591/ijece.v14i6.pp6360-6372

Crossref Full Text | Google Scholar

13. Rasa SM, Islam MM, Talukder MA, Uddin MA, Khalid M, Kazi M, et al. Brain tumor classification using fine-tuned transfer learning models on magnetic resonance imaging (MRI) images. Digital Health. (2024) 10. doi: 10.1177/20552076241286140

PubMed Abstract | Crossref Full Text | Google Scholar

14. Padmakala S and Maheswari SU. (2024). Deep Learning-based MRI Analysis for Brain Tumor Detection: Insights from ResNet and DenseNet Models, in: 2020 4th International Conference on Electronics, Communication and Aerospace Technology (ICECA), . pp. 1620–6. doi: 10.1109/iceca63461.2024.10800894

Crossref Full Text | Google Scholar

15. Kowalski M, Abdulrazak LF, Kibria HB, Khandakar A, Ayari MA, Ahamed MF, et al. A hybrid explainable model based on advanced machine learning and deep learning models for classifying brain tumors using MRI images. Sci Rep. (2025) 15(1):24778. doi: 10.1038/s41598-025-85874-7

PubMed Abstract | Crossref Full Text | Google Scholar

16. Ullah MS, Nazeeruddin E, Khan MA, and Rubab S. (2024). Brain tumor classification from MRI scans using a custom 4-residual deep learning architecture and particle swarm optimization, in: 2024 6th International Symposium on Advanced Electrical and Communication Technologies (ISAECT), Tlemcen, Algeria: IEEE. pp. 1–6. doi: 10.1109/isaect64333.2024.10799805

Crossref Full Text | Google Scholar

17. Gasmi K, Aoun NB, Alsalem K, Ltaifa IB, Alrashdi I, Ammar LB, et al. Enhanced brain tumor diagnosis using combined deep learning models and weight selection technique. Front Neuroinformatics. (2024) 18:1444650. doi: 10.3389/fninf.2024.1444650

PubMed Abstract | Crossref Full Text | Google Scholar

18. Gu C and Ren S. Enhancing brain cancer type prediction through machine learning algorithms and feature selection techniques. J Phys A Math Theor. (2024) 57(42):425601. doi: 10.1088/1751-8121/ad7e63

Crossref Full Text | Google Scholar

19. Abdfilminaam DS, Alfarouk SAO, Fouad KOM, Slait RA, and Wasfy RS. (2024). Optimizing brain tumor detection: Enhancing diagnostic accuracy in brain tumor detection using a hybrid approach of machine learning and deep learning models, in: 2024 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), Cairo, Egypt: IEEE. pp. 1–8. doi: 10.1109/miucc62295.2024.10783526

Crossref Full Text | Google Scholar

20. Li Z and Dib O. Empowering Brain Tumor Diagnosis through Explainable Deep Learning. Mach Learn Knowledge Extraction. (2024) 6(4):2248–81. doi: 10.3390/make6040111

Crossref Full Text | Google Scholar

21. Patro SGK, Govil N, Saxena S, Mishra BK, Zamani AT, Miled AB, et al. Brain tumor classification using an ensemble of deep learning techniques. IEEE Access. (2024) 12:171658–72. doi: 10.1109/access.2024.3485895

Crossref Full Text | Google Scholar

22. Regan M and Srinivasan PS. (2024). Enhancing brain tumor diagnosis with substructure aware graph neural networks and fuzzy linguistic segmentation, in: Proceedings of the 2024 International Conference on Computational Intelligence and Computing Innovations (ICOCICI), . pp. 1613–8. doi: 10.1109/icoici62503.2024.10696691

Crossref Full Text | Google Scholar

23. de Verdier MC, Saluja R, Gagnon L, LaBella D, Baid U, Tahon NH, et al. The 2024 Brain Tumor Segmentation (BRATS) challenge: Glioma Segmentation on Post-treatment MRI. arXiv (Cornell University). (2024). doi: 10.48550/arxiv.2405.18368

Crossref Full Text | Google Scholar

24. Alshammari A. Construction of VGG16 Convolution Neural Network (VGG16_CNN) Classifier with NestNet-Based Segmentation Paradigm for Brain Metastasis Classification. Sensors. (2022) 22(20):8076. doi: 10.3390/s22208076

PubMed Abstract | Crossref Full Text | Google Scholar

25. Wen L, Li X, Li X, and Gao L. (2019). A new transfer learning based on VGG-19 network for fault diagnosis, in: 2019 IEEE 23rd international conference on computer supported cooperative work in design (CSCWD), IEEE. pp. 205–9. doi: 10.1109/cscwd.2019.8791884

Crossref Full Text | Google Scholar

26. Wen L, Li X, and Gao L. A transfer convolutional neural network for fault diagnosis based on ResNet-50. Neural Computing Appl. (2020) 32(10):6111–24. doi: 10.1007/s00521-019-04097-w

Crossref Full Text | Google Scholar

27. ZainEldin H, Gamel SA, El-Kenawy EM, Alharbi AH, Khafaga DS, Ibrahim A, et al. Brain tumor detection and classification using deep learning and Sine-Cosine Fitness Grey Wolf optimization. Bioengineering. (2022) 10(1):18. doi: 10.3390/bioengineering10010018

PubMed Abstract | Crossref Full Text | Google Scholar

28. G V, Rani VV, Ponnada S, and S J. A Hybrid EfficientNet-DbneAlexnet for brain tumor detection using MRI Images. Comput Biol Chem. (2024) 115:108279. doi: 10.1016/j.compbiolchem.2024.108279

PubMed Abstract | Crossref Full Text | Google Scholar

29. Mir M, Madhi ZS, AbdulHussein AH, Al Dulaimi MKH, Suliman M, Alkhayyat A, et al. Detection and isolation of brain tumors in cancer patients using neural network techniques in MRI images. Sci Rep. (2024) 14(1). doi: 10.1038/s41598024-68567-5

PubMed Abstract | Crossref Full Text | Google Scholar

30. Dash Y, Anand P, Deepali, Chaudhary P, Sarangi SC, and Abraham A. (2024). Enhancing brain tumor classification and segmentation using ResNet50, in: Proceedings of the 2024 13th International Conference on Computing Communication and Networking Technologies (ICCCNT), . pp. 1–5. doi: 10.1109/icccnt61001.2024.10724031

Crossref Full Text | Google Scholar

31. Sadad T, Rehman A, Munir A, Saba T, Tariq U, Ayesha N, et al. Brain tumor detection and multi-classification using advanced deep learning techniques. Microscopy Res Technique. (2021) 84(6):1296–308. doi: 10.1002/jemt.23688

PubMed Abstract | Crossref Full Text | Google Scholar

32. LaBella D, Schumacher K, Mix M, Leu K, McBurney-Lin S, Nedelec P, et al. Brain Tumor Segmentation (BRATS) Challenge 2024: Meningioma Radiotherapy Planning Automated Segmentation. arXiv (Cornell University). (2024). doi: 10.48550/arxiv.2405.18383

Crossref Full Text | Google Scholar

33. Khekare G, Majumder G, Yenurkar G, Sakarkar G, Dhanre U, and Vidhale B. Identification of Brain Tumors in MRI Images Using Artificial Intelligence of Things. In: AIoT Innovations in Digital Health. CRC Press (2025). p. 93–104. doi: 10.1201/9781003469896-7

Crossref Full Text | Google Scholar

34. Gull S, Akbar S, Hassan SA, Rehman A, and Sadad T. Automated brain tumor segmentation and classification through MRI images. In: Communications in computer and information science (2022). p. 182–94. doi: 10.1007/978-3-030-97255-4_13

Crossref Full Text | Google Scholar

35. Khekare G, Bramhane LK, Dhule C, Agrawal R, and Turukmane AV. Testing and analysis of predictive capabilities of machine learning algorithms. In: Studies in computational intelligence (2022). p. 419–42. doi: 10.1007/978-3-030-99079-4_16

Crossref Full Text | Google Scholar

36. Khekare G, Turukmane AV, Dhule C, Sharma P, and Bramhane LK. Experimental performance analysis of machine learning algorithms. In: Lecture notes in electrical engineering (2022). p. 1041–52. doi: 10.1007/978-981-19-2456-9_104

Crossref Full Text | Google Scholar

37. Tan M and Le QV. EfficientNet: Rethinking model scaling for convolutional neural networks. arXiv (Cornell University). (2019). doi: 10.48550/arxiv.1905.11946

Crossref Full Text | Google Scholar

38. Campos-Delgado DU, Gutierrez-Navarro O, Rico-Jimenez JJ, Duran-Sierra E, Fabelo H, Ortega S, et al. Extended Blind End-Member and abundance extraction for biomedical imaging applications. IEEE Access. (2019) 7:178539–52. doi: 10.1109/access.2019.2958985

PubMed Abstract | Crossref Full Text | Google Scholar

39. Heyat MBB, Adhikari D, Akhtar F, Parveen S, Zeeshan HM, Ullah H, et al. Intelligent Internet of Medical Things for Depression: current advancements, challenges, and trends. Int J Intelligent Syst. (2025) 2025(1). doi: 10.1155/int/6801530

Crossref Full Text | Google Scholar

40. Iqbal MS, et al. Progress and trends in neurological disorders research. Computerized Med Imaging Graphics. (2024) 116:102400. doi: 10.1016/j.compmedimag.2024.102400

PubMed Abstract | Crossref Full Text | Google Scholar

41. Mehrotra R, Ansari MA, Agrawal R, Heyat MBB, Tripathi P, Sayeed E, et al. Deep convolutional Network-Based probabilistic selection approach for multiclassification of brain tumors using magnetic resonance imaging. Int J Intelligent Syst. (2025) 2025(1). doi: 10.1155/int/6914757

Crossref Full Text | Google Scholar

42. Vinod DS, Prakash SPS, AlSalman H, Muaad AY, and Heyat MBB. Ensemble Technique for Brain Tumor Patient Survival Prediction. IEEE Access. (2024) 12:19285–98. doi: 10.1109/access.2024.3360086

Crossref Full Text | Google Scholar

43. Abir TA, et al. Analysis of a novel MRI based brain tumour classification using probabilistic neural network (PNN). Int J Sci Res Sci Eng Technol. (2018) 4.8:65–79. https://ijsrset.com/home/issue/view/article.php?id=IJSRSET184814.

Google Scholar

44. Wickramasinghe W. M. S. P. B. and Dissanayake MB. Vision transformers for glioma classification using T1 magnetic resonance imaging. Artif Intell Health. (2024) 2:68–80. doi: 10.36922/aih.4155

Crossref Full Text | Google Scholar

Keywords: brain tumor, classification accuracy, machine learning techniques, high-grade gliomas (HGG), comparative study

Citation: Wang S and Li M (2025) Comparative analysis of machine learning techniques on the BraTS dataset for brain tumor classification. Front. Oncol. 15:1596718. doi: 10.3389/fonc.2025.1596718

Received: 20 March 2025; Accepted: 19 August 2025;
Published: 15 September 2025.

Edited by:

Sharon R. Pine, University of Colorado Anschutz Medical Campus, United States

Reviewed by:

Md Belal Bin Heyat, Westlake University, China
Rajeshkannan S, St Joseph’s College of Engineering, India

Copyright © 2025 Wang and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Min Li, bG05NzkzMDgzNjdAMTYzLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.