- 1Department of Software Engineering, University of Engineering and Technology-Taxila, Taxila, Pakistan
- 2Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
Introduction: Lung cancer remains one of the primary causes of cancer-related deaths globally, emphasizing the urgent need for accurate and early diagnosis to improve patient outcomes. However, existing computer-aided detection systems often struggle with suboptimal feature extraction, low classification accuracy, and limited generalizability across datasets.
Methods: To address these challenges, we propose a deep learning approach named Residual-SwishNet, explicitly designed for the lung cancer classification task. More specifically, we modified the ResNet50 framework by replacing the conventional ReLU activation function with Swish during the feature engineering phase. Further, we integrate three additional dense layers before the classification module to obtain an enriched feature representation. Lastly, we employ a Softmax output layer with Cross-Entropy Loss to tackle the class-imbalance issue.
Results: The approach was rigorously evaluated on 2 publicly accessible datasets, named LUNA16 and IQOTH/NCCD, using precision, recall, F1-score, and accuracy as performance metrics. Experimental results demonstrate the superiority of our technique, achieving classification accuracies of 99.60% and 99.11% on the LUNA16 and IQ-OTH/NCCD datasets.
Discussion: Our approach has significantly outperformed existing state-of-the-art techniques. These findings highlight the potential of the proposed model as a robust and reliable tool for lung cancer diagnosis.
1 Introduction
Lung cancer remains one of the most life-threatening health challenges worldwide, which originates from the abnormal and uncontrolled proliferation of cells within the lungs (1). This cancer is mainly categorized into 2 major categories: non-small cell lung cancer (NSCLC) and small cell lung cancer (SCLC) (2). Among these, NSCLCs are common and usually develop at a slower rate than SCLCs. Globally, lung cancer continues to be one of the most frequently identified tumors and a major cause of death. According to reports by the World Health Organization (WHO) and the International Agency for Research on Cancer (IARC), this category of cancer ranked as the second most diagnosed tumor in 2021, having 2.2 million new cases, exceeded only by breast cancer. Frighteningly, it was responsible for approximately 1.8 million deaths out of a total of 9.96 million cancer-related mortalities (3). The trend continued in 2022 and resulted in affecting nearly 2.5 million individuals and causing the same number of deaths, supporting its deadly impact (4). These alarming figures reflect the urgent need for more accurate, efficient, and early diagnostic techniques for this deadly disease. Timely recognition plays a pivotal part in refining prognosis and survival proportions. In this context, the current research is motivated by the persuasive challenge of high mortality and diagnostic complexity associated with lung cancer. This study proposes a novel deep learning (DL)-based method for recognizing lung cancer nodules with the aim of enhancing diagnostic reliability and supporting clinical decision-making. A crucial aspect of lung cancer analysis involves the detection and assessment of lung nodules, which are small, abnormal growths that appear on imaging modalities like chest X-rays and computed tomography (CT) scans (5). These nodules are either benign or malignant, with the latter potentially indicating early stages of lung cancer. Effective evaluation of these nodules is essential, as it can result in prompt intervention and significantly improved treatment outcomes. Consequently, routine screening, accurate classification, and conscientious monitoring of lung nodules form the cornerstone of modern strategies aimed at reducing the burden of lung cancer and improving patient care (6).
CT scans stand out as one of the most effective tools among the numerous imaging modalities used in the recognition of lung cancer. CT imaging provides a detailed cross-sectional representation of the chest, which permits clinicians to closely examine the lungs for signs of malignancy or other abnormalities (7). Its aptitude to produce high-quality, three-dimensional images promptly makes it a vital resource for the timely detection, recognition, and staging of lung tumors. By offering enhanced visual clarity, CT scans play a critical role in differentiating between benign and malignant nodules, supporting accurate clinical assessments (8). Additionally, CT imaging serves as a guide for subsequent diagnostic measures, like needle biopsies, and aids in devising appropriate cure strategies and observing disease progression (9). Despite its diagnostic value, the traditional manual interpretation of CT scans poses several challenges. Reviewing a large volume of medical samples is laborious and mentally taxing for experts. This manual process is not only labor-intensive but also prone to inconsistencies, as complex anomalies can go unnoticed or be interpreted differently by different observers. The inherent subjectivity in image analysis eventually results in inter-observer variability, which compromises diagnostic accuracy. Furthermore, with the increasing availability and use of advanced imaging technologies, the sheer amount of data generated has placed an added burden on healthcare professionals. Such limitations demand the requirement for computerized, intelligent systems that can assist in the accurate and efficient interpretation of lung imaging data (10).
Recent advancements in automated lung cancer detection have increasingly relied on imaging data across several modalities, mostly through the application of artificial intelligence (AI) and machine learning (ML) methods. These approaches have appeared as powerful solutions for analyzing complex medical images, which empower them to identify the complicated patterns that are hard to detect through traditional methods (11). AI-driven approaches, especially when applied to CT scans, are capable of efficiently processing huge amounts of imaging data, which supports radiologists in pinpointing minute anomalies associated with lung tumors or nodules. This capability significantly enhances both the speed and precision of diagnosis. By automating the analysis process, AI not only accelerates early detection efforts but also plays a crucial role in standardizing diagnostic practices across diverse clinical environments. It reduces reliance on subjective interpretation, thus minimizing variability between observers and enhancing consistency in results. The integration of intelligent diagnostic systems supports healthcare specialists in making accurate and timely decisions, eventually causing better patient management and improved treatment outcomes (12). These innovations underscore the transformative influence of AI in modern medical imaging and reinforce its value in the fight against lung cancer (13).
Early efforts in lung cancer detection primarily relied on traditional ML algorithms, which involved analyzing hand-crafted features computed from medical samples (14). Investigators focused on quantifiable attributes like the outline, texture, and concentration of suspected lesions in CT scans to perceive and categorize irregularities. These features were then passed to train classifiers that aimed to discriminate between the benign and malignant cases. While this approach yielded encouraging results, it was constrained by the challenges associated with medical image complexity (15). One of the key limitations of traditional ML methods was their dependence on manual feature extraction, which required significant field knowledge and often failed to capture delicate or higher-order patterns within the data. Furthermore, conventional ML techniques struggled to manage the hierarchical and nonlinear characteristics of medical images, which limit their capability to generalize across diverse patient datasets. As a result, their performance in the timely and precise recognition of lung cancer was suboptimal. The advent of DL revolutionized the landscape of medical image examination, mainly in the area of lung cancer classification (16). DL architectures like convolutional neural networks (CNNs) enabled models to automatically compute hierarchical and complicated features right from a given suspected sample (17). Such architectures eradicate the requirements for a manual feature learning procedure and significantly improve the ability to compute fine-grained and abstract patterns associated with cancerous growths (18–20). DL-based models have demonstrated exceptional performance in recognizing sophisticated visual cues, for example, small nodules or early-stage tumors, often undetectable to the human eye. The end-to-end learning capability of deep networks allows them to process raw data and optimize predictions with minimal human intervention, thereby progressing the correctness, consistency, and reliability of lung cancer diagnostics (21–23).
Although DL frameworks have meaningfully advanced lung cancer detection, they are not without limitations. A major concern is the risk of diagnostic errors, including false positives where non-cancerous structures are misidentified as malignant and false negatives, where actual tumors may go undetected. These misclassifications can have serious clinical implications, potentially leading to unnecessary interventions or missed treatment opportunities. Another persistent challenge is the lack of interpretability in DL models. Many DNNs operate as opaque systems, making it problematic for medical experts to understand the reasoning behind a specific prediction, thereby reducing trust and limiting their integration into decision-critical environments. In addition, the computational demands of DL models are substantial. Training and deploying these models typically require high-performance hardware and significant processing time, which can be a barrier to real-time application in clinical workflows, particularly in resource-constrained sites. Furthermore, the outcomes of DL algorithms can be influenced by disparities in imaging protocols, scanner types, and image quality. These inconsistencies across datasets and healthcare institutions may hinder the generalizability of the models, limiting their robustness and effectiveness when applied to new or diverse populations. Addressing these limitations is essential to ensure the safe, efficient, and wide-scale adoption of DL systems in lung tumor recognition and classification.
To address the existing challenges of this field, we propose a refined ResNet50-based model named Residual-SwishNet, designed to enhance classification performance and model robustness. In our approach, the conventional activation functions in the convolutional layers have been replaced with the Swish activation function, known for its smooth, non-monotonic properties that enable improved gradient propagation and feature computation. This modification permits the network to extract complicated patterns more effectually, specifically in complex or low-contrast regions of lung scans. Additionally, we introduce three extra dense layers before the final categorization layer to enrich the feature space and improve the model’s capacity to recognize benign and malignant samples. Further, the usage of the cross-entropy loss method in the softmax layer assists the Residual-SwishNet approach to tackle the class-imbalance problem. These architectural enhancements contribute to more reliable predictions and help mitigate false positives and negatives by offering a more accurate and interpretable solution for lung cancer classification. The main contributions of this framework are listed as:
1. Integration of the Swish activation function in convolutional layers improves non-linear representation learning and model convergence, outperforming traditional functions like ReLU.
2. The addition of three fully connected layers before the classification unit enhances the capability of the approach to extract deep semantic features, which leads to improved decision boundaries.
3. The proposed architecture demonstrates better handling of ambiguous cases due to its high recall ability, which minimizes the incidence of false positives and negatives in lung cancer classification.
4. The model is capable of tackling the class-imbalance issue due to the employment of the cross-entropy loss method in the softmax layer of the proposed approach.
5. The refined model offers a practical and computationally efficient solution that bridges the gap between high accuracy and interpretability, which makes it suitable for deployment in clinical diagnostic workflows.
6. Massive experimental analysis of our work was performed using two standard datasets to demonstrate the efficacy of our method.
The rest of the article comprises the given hierarchy: the related work is provided in Section 2, whereas the suggested work is presented in Section 3. The used dataset, performance indicators, along with a detailed description of achieved outcomes, are given in Section 4. Lastly, the conclusion is stated in Section 5.
2 Related work
Here, we have discussed the existing works presented to accomplish the classification task of lung nodules.
Raza et al. (24) proposed a model named Lung-EffNet for classifying lung cancers. Initially, the approach performed data augmentation to increase the sample size. After this, the EfficientNet approach was trained using the concept of transfer learning via applying its variants from B0 to B4. The work was tested over the IQ-OTH/NCCD dataset and obtained a classification score of 99.10%. This work shows good outcomes for categorizing lung nodules; however, it requires a huge number of samples to train the model. Nahiduzzaman et al. (25) introduced a DL method for classifying lung tumors. For this, the model first applied a preprocessing step to enhance the visual representation of the input images. After this, the work applied a model named lightweight parallel depth-wise separable convolutional neural network (LPDCNN) to compute the relevant information of the input images. Next, the Ridge-ELM classifier employed the extracted features from the previous step and performed the categorization of nodules. The work provides better results for binary problem; however, it faces performance degradation for the multiclass classification task. Another DL approach named FocalNeXt was presented in (26) to execute the classification task of lung nodules. The work merged the attention strategy of FocalNet along with the feature computation capability of ConvNeXt inside the vision transformer model. The work has attained a classification score of 99.81%, however, at the cost of an enhanced computing load. Priya et al. (27) proposed a DL model for lung cancer classification. First, samples were preprocessed to improve their appearance. Then, the data size was increased by applying various augmentation techniques. Finally, the work used the SE-ResNeXt-50 approach to calculate the deep features and performed the classification job. The work performed well in recognizing lung cancers; however, it suffered from a high computational cost.
In (28), a DL method was used to recognize the lung tumors. First, a step was followed to eliminate the noisy information from the samples and enhance their visual appearance. The enhanced samples were then passed to the feature extraction phase, where 6 statistical features were estimated based on Improved Empirical Wavelet Transforms (IEWT). After this, the calculated visual characteristics were passed to the Attention-based Convolutional Neural Network with DenseNet-201Transfer Learning (AtCNN-DenseNet-201 TL) to accomplish the categorization task. The approach presents an effective solution to diagnose lung cancer; however, it faces the model overfitting issue. Murthy et al. (29) used an ML approach to categorize lung cancer from CT scans. For this, the Adaptive Median Filter (AMF) was used initially to enhance the quality of the samples. Next, M-SegNet was used to extract the focal areas. After this, the visual information from the segmented regions was computed by computing various types of features named statistical, deep, and textural features, which were combined to make the final vector. After this, the classification was accomplished using the Tree-based Pipeline Optimization Tool with Support Vector Machine (TPOT_SVM). The approach shows better outcomes in the recognition of lung nodules; however, the work is effective for a binary classification task. Mothkura et al. (30) designed a method for classifying lung tumors. The approach used a dense CNN to get dense features and performed the classification task. The work has attained a highest classification rate of 85.21%, which needs further improvements. In (31), a DL network was proposed for lung cancer classification. To accomplish this, the work used s Multimodal Fusion Deep Neural Network (MFDNN) framework that combined features from various types of visual samples like medical imaging, genomics, and clinical data, which was combined for lung cancer classification. The work performed well in recognizing the lung cancer nodules; however, at the cost of a huge computing burden. Mohamed et al. (32) suggested a CNN approach to extract a related group of features and perform the categorization of samples into relevant groups of lung cancers. Further, the work used an optimization strategy to boost the model’s running behavior. The work shows better outcomes in recognizing the relevant cancer groups of the lung; however, the results need further enhancements.
Hossain et al. (33) employed the concept of the ensemble approach for categorizing lung cancer nodules. For this, the approach initially utilized a step to boost the quality of input samples. After this, the work used the CNN-SVD-Ensemble approach to compute the relevant set of visual information and execute the dimensionality reduction. Next, the computed features were passed to various ML classifiers to perform the cataloguing of lung nodules. The strategy gains better outputs in diagnosing the lung tumors; however, it lacks the generalization power. Naseer et al. (34) proposed a DL approach to perform the lung cancer classification task. For this, the approach follows 3 stages, where in the first step, the work used the UNet approach to segment the lobe from CT scans. Next, the enhanced UNet framework was used to compute the nodule mask. In the third stage, the computed masks were passed to a CNN approach named the AlexNet for deep features computation and executing the classification task. The network was tested using the LUNA-16 data sample and attained an accuracy score of 97.98%. The technique attains better scores; however, it requires more improvements. Uddin et al. (35) applied a DL network for lung cancer diagnosis. For this, the work proposed two dense frameworks to learn a representative set of features and execute the recognition task. Further, the work has proposed a hybrid model by joining both networks, and attained a highest classification accuracy of 93%. This technique exhibits better performance in categorizing lung cancer nodules; however, the outcomes need further improvement. In (36), a DL network was designed for the recognition of lung tumors. For this, initially, the samples were preprocessed to enhance the visual information by propagating the sample to the convolution filter and down-sampling it by employing max pooling. Next, the autoencoder approach was applied to compute a set of dense features based on a CNN, and a multispace image reconstruction method was utilized to lessen error while restructuring the sample to enhance the nodule recognition ability of the approach. Lastly, the SoftMax classifier was applied for grouping the samples. Alsallal et al. (37) designed a framework to identify and classify lung tumors. First, various transformations were used to advance the visual aspects of the samples. Next, various structural features were computed by applying the PyRadiomics library. Next, a dense CNN model was merged with attention mechanisms to compute a deep set of sample information. Next, the model used different feature nomination approaches like Non-negative Matrix Factorization (NMF) and Recursive Feature Elimination (RFE). Finally, various ML classifiers like XGBoost and Stacking were applied to perform the classification task. The work performs well for lung cancer classification; however, it needs further improvements.
While numerous studies have explored ML and DL methods for recognizing lung tumor nodules, several limitations remain unresolved. Many existing approaches struggle with issues such as limited generalizability, lack of interpretability, and susceptibility to false predictions. These gaps underline the pressing need for a more effective and robust solution that can enhance feature representation, improve diagnostic accuracy, and ensure consistency across diverse imaging conditions.
3 Proposed method
This study presents a novel DL model named Residual-SwishNet, which is specifically designed to classify lung CT images into benign and malignant groups accurately. The proposed architecture builds upon a refined version of the ResNet50 CNN model, which is modified to capture relevant and complex details found in lung cancer imaging. To improve the learning capability of the approach, we replace the conventional ReLU activation with the Swish activation within the convolutional layers. The smooth and non-monotonic behavior of the Swish activation allows for better gradient flow and the retention of minor negative values, which enhances feature learning in deep layers. Furthermore, the architecture is extended by introducing three additional dense layers before the final classification stage. These layers strengthen the capability of the technique to extract deep semantic features and support more precise decision-making. This enriched feature representation contributes to more reliable classification outcomes, particularly in cases where tumor boundaries or nodular characteristics are complicated. Moreover, to tackle the category imbalance issue, we have used the cross-entropy loss method in the final classification stage. A workflow of the model training is provided in Figure 1. For model evaluation, the employed datasets were partitioned into train and test groups, and framework parameters were optimized to ensure stable convergence and robust performance. Residual-SwishNet was then trained to identify key discriminative features in lung CT scans and accurately categorize them into their respective classes. A representation of the designed framework is given in Figure 2.
3.1 Feature calculation
To boost the accuracy of lung cancer classification, it is essential to compute a robust set of feature representations that effectively capture the differences between malignant and benign lung tissues. High-level features, often referred to as semantic features, are vital in distinguishing such complex visual patterns in CT scan images. In the proposed Residual-SwishNet framework, we utilize a modified ResNet50 backbone to extract these discriminative features. A key enhancement involves replacing the traditional ReLU activation with the Swish, which advances the aptitude of our framework to compute complex spatial information by permitting small negative scores to pass through the network. This helps retain valuable information during feature learning and supports deeper, more expressive representations. As a result, the feature computation stage in Residual-SwishNet becomes more effective in encoding semantic details crucial for improving classification performance.
3.2 ResNet50
The base of the proposed Residual-SwishNet framework lies in the ResNet50 architecture (38), which is a powerful and widely adopted deep CNN model known for its use of residual learning. Unlike traditional CNNs that suffer from vanishing gradients as layers deepen, the ResNet50 model addresses this issue through identity-based skip links, which allow gradients to flow more successfully through backpropagation. These skip links empower the system to capture residual mappings rather than direct transformations, which results in both improved convergence and accuracy in deep architectures.
ResNet50 is structured using multiple residual blocks (RBs), each consisting of a series of convolutional layers accompanied by batch normalization and an activation method. The distinguishing feature of an RB is the shortcut path, which allows the input to avoid one or more layers and be directly added to the output of the transformation layers. This process is mathematically represented as Equation 1.
Where o is the input to the block, R(o) is the residual mapping learned through convolutional operations, and I is the final output. This formulation helps hold vital features across layers while enabling the approach to focus on learning only the necessary adjustments. Owing to its architectural efficiency and strong feature extraction capabilities, the ResNet50 model serves as an effective base for the Residual-SwishNet model in analyzing lung CT images for cancer classification.
3.3 Residual-SwishNet
The proposed Residual-SwishNet framework builds upon the ResNet50 architecture (39), which is famous for its residual learning mechanism and robust feature extraction capabilities. Even though ResNet50 performs well in solving various image classification tasks, however comes with certain limitations when applied directly to medical imaging, mainly for detecting complex abnormalities in lung CT scans. Specifically, its dependence on the ReLU activation function leads to the suppression of important negative activations, which can carry useful information in low-contrast medical images. Additionally, the standard architecture is unable to provide sufficient depth in the final layers to fully capture the complicated variations between benign and malignant tissues. To address these challenges, we introduce three key modifications to the original ResNet50 structure, which build the foundation of the Residual-SwishNet model. First, we incorporate the Swish activation function into the residual blocks to replace ReLU in the process of feature capturing. The capability of the Swish method to retain small negative values and promote smoother gradient flow enhances the power of the model to learn complex and refined visual features by improving convergence and reducing training loss. Second, we append 3 extra dense layers before the classification unit to deepen the semantic understanding behavior of the model, which improves its discrimination power before executing the final grouping task. These fully connected layers refine the high-level features and help the model detect more meaningful patterns associated with lung abnormalities. Lastly, we have used the cross-entropy loss technique in the classification layer, which assists the Residual-SwishNet in tackling the sample imbalance problem. Despite these enhancements, the overall complexity of the network remains manageable, which ensures that Residual-SwishNet is suitable for practical clinical applications. The final feature representations are passed to a Softmax layer, which outputs class probabilities for the classification of lung conditions. The architectural overview of Residual-SwishNet is illustrated in Figure 3, and a detailed layer-wise structure is provided in Table 1. The next sections delve into the internal components of the model in greater detail.
3.3.1 Convolutional layer
In the Residual-SwishNet architecture, convolutional layers form the foundational component responsible for extracting spatially-aware features from lung CT scan images. These layers apply a series of learnable filters to the input image, which enables the technique to detect local patterns such as edges, textures, and shapes that are critical for identifying cancerous regions. As the depth of the network increases, these layers progressively learn more abstract and complex representations, which are essential for distinguishing between benign and malignant conditions. -The operation of a convolutional layer can be mathematically expressed as Equation 2.
Where denotes the output feature map at layer n, represents the convolutional kernel applied between input map i and output map j, ∗ is the convolution operator, is the bias term, f(·) denotes the activation function (Swish in our case), and denotes the total features maps. To standardize the input across all samples and maintain consistency with the architecture, CT scan images are resized to 224 × 224 pixels before being passed through the network. In total, the proposed Residual-SwishNet model incorporates 48 convolutional layers, which are distributed across multiple residual blocks, and each contributes to the hierarchical extraction of features necessary for accurate lung cancer classification.
3.3.2 Activation layer
To enhance the capacity of the approach to learn fine-grained features from lung CT images, the proposed Residual-SwishNet architecture replaces the conventional ReLU activation with the more advanced Swish activation function after every 2D conv layer. Swish is famous for its smooth and non-linear behavior that enables better gradient propagation and feature retention, particularly in deeper networks. Its ability to pass small negative values in place of completely discarding them, as ReLU does, assists the proposed work in preserving complex patterns in medical images that might otherwise be lost, which is especially critical for detecting early-stage or low-contrast lung abnormalities (40). Unlike ReLU, which is piecewise linear and can lead to dead neuron problems, the employed Swish activation function offers a non-monotonic and differentiable structure, enabling better optimization and more stable training. This leads to stronger generalization performance and improved convergence behavior. A pictorial analysis of the Swish vs ReLU methods is depicted in Figure 4. The mathematical expression for the Swish activation is given by Equation 3.
Figure 4. Pictorial analysis of ReLU and Swish activation (40).
Where u is the input to the activation layer, σ denotes the sigmoid function, and η is a trainable parameter that scales the input dynamically. However, in our implementation, the parameter η in the Swish equation is fixed to 1 (non-trainable), corresponding to the standard Swish activation rather than the trainable Swish-β variant. The fixed version provided more stable convergence and avoided additional parameters that could lead to overfitting on medical datasets with limited samples. In the context of lung cancer classification, in which learning relevant morphological changes is crucial, the Swish activation approach significantly improves feature sensitivity and learning depth. Its computational efficiency and stability under varying learning rates make it particularly suitable for deep architectures like Residual-SwishNet, which delivers better performance and faster convergence compared to traditional activation functions.
3.3.3 Pooling layer
In the Residual-SwishNet architecture, pooling layers are integrated to reduce the spatial dimensions of feature maps while retaining the most informative patterns. The pooling operation helps in minimizing redundancy, improving computational efficiency, and enhancing the ability of the model to manage spatial variations in lung CT scans by aggregating features from neighboring pixels. Specifically, an average pooling layer is employed before the fully connected stages to summarize the extracted features into a compact vector, which is then passed to the appended dense layers for further processing and classification.
3.4 Incorporated dense layers
After the pooling stage, Residual-SwishNet incorporates 3 additional dense layers to refine and enhance the extracted features before classification. The decision to incorporate three dense layers in the final stage of Residual-SwishNet was guided by empirical experimentation conducted during model development. We evaluated multiple configurations, including one, two, three, and more than three dense layers. Models with only one or two dense layers demonstrated weaker feature abstraction and produced lower classification accuracy, indicating insufficient refinement of high-level representations. In contrast, architectures with more than three dense layers began to show signs of overfitting and increased training instability. The configuration with three dense layers consistently achieved the most favorable balance between feature discrimination, computational efficiency, and generalization capability. These layers are activated using ReLU and are focused on emphasizing critical patterns associated with lung abnormalities while filtering out less relevant information. This added depth improves the capacity of the approach to form a more discriminative feature representation from CT scans. The resulting dense feature vector is then forwarded to the Softmax layer for final classification.
3.5 Softmax layer
The last stage of the Residual-SwishNet framework utilizes a Softmax layer to convert the dense feature vector into class probabilities, which causes the model to classify lung CT images as benign or malignant. The Softmax function normalizes the outputs of the last dense layer by producing a probability distribution over the target classes. It is defined as Equation 4.
where is the score for class i, presents the output vectors, and n shows the total groups. For optimization, the model uses categorical cross-entropy loss (41), which measures the divergence between predicted and true class labels and supports learning even under class imbalance.
4 Results
In this section, a thorough explanation of the employed datasets, along with the evaluation parameters, is provided. Further, a vast evaluation of the Residual-SwishNet is performed to prove the robustness of our network.
4.1 Implementation details
The proposed Residual-SwishNet was implemented using Python and TensorFlow/Keras. The model was optimized using the Adam optimizer with an initial learning rate of 0.0001, chosen for stable convergence when training deep residual networks. A batch size of 32 was used for all experiments to maintain a balance between training stability and computational efficiency. Both datasets were divided into 70% training, 10% validation, and 20% testing splits to enable fair performance evaluation. The model was trained for 100 epochs with early stopping based on validation loss. Further, Figure 5 shows the training and validation curves for both datasets. As the epochs increase, the accuracy for both training and validation steadily improves, while the loss gradually decreases. The two curves stay close to each other, which means the model is learning well and not overfitting. These curves confirm that our Residual-SwishNet trains smoothly and generalizes well to unseen data.
Figure 5. LUNA-16 Training curves (a), LUNA-16 Loss curves (b), IQ-OTH/NCCD Training curves (c), IQ-OTH/NCCD Loss Curves (d), respectively.
4.2 Evaluation parameters
To measure the classification outcomes of the proposed Residual-SwishNet technique, we have used the standard evaluation parameters named accuracy, precision, recall, and F1-Score, which are computed as mentioned in Equations 5–8.
4.3 Dataset
To train and test the proposed approach, we utilized two standard online repositories of lung cancer samples, namely LUNA-16 (42) and IQ-OTH/NCCD lung cancer dataset (43), both of which were accessed on July 01, 2025. The LUNA16 dataset, a curated subset of the publicly available LIDC-IDRI repository, is a broadly recognized standard for tuning networks for lung cancer diagnosis. It contains chest CT scans that have been meticulously annotated by expert radiologists, who evaluated each scan to confirm the incidence, locality, and characteristics of pulmonary nodules. As part of its refinement, the LUNA16 dataset excludes CT images with slice thicknesses greater than 2.5 mm to ensure consistent imaging quality across the dataset. The complete dataset comprises 888 CT scans, which are reviewed by four radiologists, providing reliable annotations of the size, location, shape, and density of nodules. These images focus specifically on the lung regions containing nodules, allowing the framework to extract appropriate patterns. To facilitate the lung tumors classification task, the nodules were divided into two categories based on the mean malignancy score assigned by radiologists. Nodules with a mean score of 2.5 or below were categorized as benign, indicating non-cancerous conditions, while those with a mean score of 3.5 or higher were classified as malignant, reflecting a higher likelihood of cancer. The second data sample utilized in this work is the IQ-OTH/NCCD lung cancer dataset, which was gathered at the National Center for Cancer Diseases in Iraq over a period of 3 months during the fall of 2019. This sample comprises a total of 1,190 CT image slices obtained from 110 victims and delivers an adequate amount of data for framework tuning. The dataset is organized into three distinct categories: normal, benign, and malignant cases. Among the collected samples, there are 416 images labeled as normal, 120 as benign, and 561 as malignant, which indicates a minor imbalance in class distribution. The major reason to choose these datasets is due to their complex natures and presence of various image distortions, which make them challenging for accomplishing the lung cancer classification task.
4.4 Model evaluation
This section is focused on discussing the output results of the Residual-SwishNet approach by indicating the attained scores on two employed standard datasets named LUNA16 and IQ-OTH/NCCD lung cancer data samples, with the employment of various measures like precision, recall, F1, and Accuracy metrics.
To thoroughly assess the results of our proposed Residual-SwishNet model, first, we discussed the classification metrics named precision, recall, and F1-score on both the LUNA16 and IQ-OTH/NCCD samples, and attained analysis is provided in Figure 6. These metrics offer a complete understanding of the diagnostic strength and reliability of our model, which is vital in health imaging tasks where both false positives and negatives can possess serious clinical consequences. Precision reflects the capability of an approach to appropriately recognize true positive cases among all predicted positives, which is vital to minimizing unnecessary interventions. Recall, on the other hand, computes the capacity of the work to detect all actual positive cases by ensuring that no malignant or suspicious case is overlooked. Lastly, the F1-score presents a harmonic mean of precision and recall, which balances these two aspects by delivering a single robust indicator of classification performance, especially under class-imbalanced conditions common in medical datasets. Using line bar visualizations, we plotted these metrics across both datasets to showcase how Residual-SwishNet consistently achieves high scores across all three metrics, showing the efficacy of our network. These outcomes confirm that our approach is competent to maintain high diagnostic accuracy while reducing both false alarms and missed detections, proving its effectiveness in real-world clinical screening circumstances.
Figure 6. Residual-SwishNet results over the LUNA16 and IQ-OTH/NCCD datasets using precision, recall, and F1-score measures.
In addition to evaluating precision, recall, and F1-score, we also analyzed the accuracy of our proposed model across both datasets, and the results are shown in Figure 7. Accuracy is one of the most fundamental metrics in classification, which computes the proportion of correctly predicted instances over the total number of predictions. On both the LUNA16 and IQ-OTH/NCCD datasets, our model consistently outperformed and attained improved accuracy values. It can be seen from the values provided in Figure 7 that our approach offers comprehensive capability to correctly classify a wide range of lung cancer cases, regardless of their complexity or similarity across classes. The improved accuracy is due to the better information-capturing capability of the approach, which ensures minimal misclassification and boosts the model’s reliability for clinical use.
Figure 7. The Residual-SwishNet results representation over the LUNA16 and IQ-OTH/NCCD datasets using accuracy metrics.
Next, we have discussed the confusion matrix for both employed datasets, and visuals are provided in Figure 8. The confusion matrices shown in Figure 8 not only reflect the exceptional performance of the proposed model on the LUNA16 and IQ-OTH/NCCD datasets but also serve as a critical tool for evaluating classification effectiveness. A confusion matrix provides a comprehensive view of the predictive capabilities of an approach by presenting true positives, true negatives, false positives, and false negatives in a structured format. This aids in analyzing explicit zones where the approach surpasses or needs development, beyond what single-value metrics like accuracy or F1-score can convey. In addition to the visual confusion matrices shown in Figure 8, we also provide the numerical confusion matrix values in tabular form to improve clarity in Tables 2, 3 for LUNA-16 and IQ-OTH/NCCD datasets, respectively. These tables present the exact number of correctly and incorrectly classified samples for each class in both datasets, enabling a more precise interpretation of the model’s performance. Specifically, for the LUNA16 dataset, the model accurately classifies 99.84% of benign and 99.79% of malignant cases, with extremely low misclassification rates of 0.16% and 0.21% respectively, indicating high discriminative power between the two classes. While for the IQ-OTH/NCCD dataset, presenting a 3-class problem named normal, benign, and malignant, the confusion matrix again demonstrates excellent performance with 99.89%, 99.85%, and 99.78% correctly predicted for each class, respectively. The extremely low off-diagonal values confirm the capability of the approach to avoid confusion between clinically similar categories. These strong results validate the effectiveness of the proposed modifications, like replacing ReLU with Swish activation, incorporating three dense layers, and using a Softmax output layer with Cross-Entropy Loss, in enhancing feature learning and decision boundaries, ultimately leading to superior classification precision and reliability.
Figure 8. Confusion matrix attained by the Residual-SwishNet over (a) LUNA16, (b) IQ-OTH/NCCD, respectively.
To qualitatively assess the model’s recognition capability, we presented a few example CT scan images marked with both the predicted and original labels in Figure 9. It can be seen from the visuals given in Figure 9 that the Residual-SwishNet accurately classifies the cases, clearly demonstrating its ability to generalize and correctly identify critical patterns in complex lung imagery. Such visual validations strengthen the confidence in our proposed solution by highlighting its potential reliability in clinical environments. The sharp correspondence between the projected and real labels confirms the robustness of the proposed architecture across diverse imaging conditions.
To further validate the robustness of Residual-SwishNet, ROC curves were generated for both datasets. As shown in Figure 10, the model achieves an AUC of approximately 0.97 on LUNA-16 and 0.98 on IQ-OTH/NCCD, indicating excellent separability between benign and malignant cases across both datasets. The curves rise sharply toward the top-left corner, demonstrating a high true-positive rate with minimal false positives, which confirms stable generalization beyond accuracy and F1-score results. These ROC curves provide additional evidence that the proposed method maintains strong discriminative performance and does not rely on chance-level decision boundaries.
Figure 10. ROC curves for both datasets (a) LUNA-16, (b) IQ-OTH/NCCD using the Residual-SwishNet approach.
To further support the interpretability of the proposed model, we have generated heatmaps using Grad-CAM, and attained visuals are given in Figure 11. These visualizations highlight the explicit areas within the CT scans that the Residual-SwishNet emphasizes when making predictions. The activation regions primarily concentrate on the interior portions of the lungs, indicating that the model is correctly attending to clinically relevant areas for feature extraction. This not only validates the internal decision-making process of the deep learning model but also aligns with expert radiological understanding, thus reinforcing the clinical reliability and transparency of the system.
So, after providing a comprehensive assessment of the Residual-SwishNet approach using multiple performance metrics, including precision, recall, F1-score, and accuracy, along with visual analysis through confusion matrices, predicted label overlays, and Grad-CAM heatmaps, we have confirmed the robustness of our approach for lung cancer classification. The consistently high scores across both datasets and the strong qualitative evidence highlight the effectiveness, robustness, and interpretability of the model in classifying lung cancer cases. These evaluations confirm that the model not only attains advanced performance but also offers transparency in its predictions, which is an essential requirement for real-world medical applications.
4.5 Comparison with DL approaches
In this part, we have analyzed the performance of the Residual-SwishNet approach in comparison to numerous well-known DL approaches for both employed datasets, named LUNA16 and IQ-OTH/NCCD lung cancer data samples.
First, the scores for the LUNA16 dataset are discussed by using the standard evaluation measures, namely precision, recall, F1-score, and accuracy, and the obtained analysis is shown in Table 4. For evaluation, various DL approaches, i.e., the EfficientNet series (44), NASNetMobile (45), DenseNet121 (46), MobileNetV2 (47), and MobileNet (48), are taken as provided in (49). The assessment in Table 4 clearly displays that the proposed solution ranked highest in all evaluation measures with scores of 99.90%, 99.81%, 99.85%, and 99.60% for the precision, recall, F1-Score, and accuracy metrics. Further, the EfficientNetB2 model attains comparable performance with scores of 95.50%, 94.60%, 95.10%, and 95.40% against the precision, recall, F1-score, and accuracy measures. Among all DL models, the NASNetMobile approach shows the worst performance in classifying the lung cancer samples with scores of 88.25%, 86.65%, 87.65%, and 87.45% over the precision, recall, F1, and accuracy metrics. The proposed solution shows the best solution due to its high recognition power. On the LUNA16 dataset, the comparative DL models show better results; however, these models are subject to certain limitations in the context of lung cancer classification. The EfficientNet and EfficientNetV2 series techniques present optimized solutions for parameter efficiency; however, at the cost of compromised accuracy due to their shortcoming in capturing the complex textural differences present in lung nodules, as these models employ lightweight convolutional layers and standard activation functions. Such a structure of the EfficientNet family limits their feature sensitivity. Further, NASNetMobile and MobileNet architectures prioritize computational efficiency for mobile applications but sacrifice depth and feature extraction capabilities, which also makes them less suitable for high-detail medical imaging tasks. DenseNet121, though effective in feature reuse, lacks specialized modifications tailored to lung cancer detection and does not optimally capture domain-specific patterns. In contrast, our Residual-SwishNet enhances feature learning through a ResNet50 backbone equipped with the Swish activation function, which improves non-linear learning and gradient flow, and integrates additional dense layers to refine feature representation. These improvements enable Residual-SwishNet to better capture fine-grained nodule characteristics, resulting in superior classification accuracy, sensitivity, and specificity compared to these general-purpose models.
Next, the comparison results for the proposed solution are discussed over the IQ-OTH/NCCD dataset using various DL models like DenseNet-121 (46), Inception-V3 (50), MobileNet-V2 (47), ResNet-50 (51), ResNet-152 (52), ResNet101 (53), as given in (54, 55), and the acquired evaluation is given in Table 5. Again, the results show that our model, named Residual-SwishNet, performs better than the selected DL approaches in classifying lung cancer over the IQ-OTH/NCCD dataset with a precision of 99.86%, along with F1-score and accuracy values of 98.85% and 99.11%. The comparative models have proven effective in generic image classification tasks; however, they exhibit key limitations in medical imaging contexts. DenseNet-121 and Inception-V3, despite their depth and feature reuse capabilities, rely on conventional activation functions and lack targeted optimization for lung cancer detection. MobileNet-V2 is a lightweight architecture that compromises feature depth for computational efficiency, which reduces its ability to extract detailed lung nodule patterns. Even the deeper ResNet variants are powerful, but employing ReLU activation and standard architectural setups, which limit their capacity to capture detailed variations critical in lung CT scans. In contrast, our Residual-SwishNet builds on ResNet50 but overcomes these limitations through the integration of the Swish activation, which advances gradient propagation and feature sensitivity. Further, introducing additional dense layers resulted in a more refined feature representation. These enhancements allow Residual-SwishNet to extract and process intricate features more effectively, resulting in improved results across all key measures on the IQ-OTH/NCCD dataset.
4.6 Comparison with the state-of-the-art
In this part, we have evaluated the designed work against various latest approaches employing the same dataset to validate its robustness in classifying the lung cancer nodules.
First, we have discussed evaluation results for the LUNA16 dataset by analyzing its outcomes with numerous new works (27, 49, 56–60). Priya et al. (27) suggested a DL framework named SE-ResNeXt-50 to classify the lung cancer nodules and attained an accuracy score of 99.15%. In (60), a DCSwinB dual-branch DL method was proposed that combines CNN-based local feature extraction with Swin Transformer-based global context modeling, fused through a Conv-MLP module for enhanced 3D representation, and attained an accuracy score of 90.96%. The approach in (49) proposed a CNN approach with an attention strategy to recognize lung nodules from CT-Scan images. The CNN unit computed deep features on which an attention module was employed to recognize the relevant information. The work reported an accuracy score of 95.40%. Thangavel et al. (56) utilized a DL model to categorize pulmonary nodules from suspected images. Initially, some preprocessing steps were performed to enhance the visual quality of samples, on which the TNet DL model was applied to segment the focal regions from the CT-Scan images. After this, the CenterNet technique was applied to compute visual aspects from the extracted clusters. At last, the NASNet approach was applied to execute the classification task. This dense network has obtained an accuracy score of 99.29%. Next, discussed work in (57) proposed a model named LungNet-SVM for lung cancer classification. The approach presented an improved AlexNet approach for deep features computation, on which the classification was carried out by the SVM classifier. The work has achieved an accuracy score of 97.64%. Next, the method in (58) applied Gabor filters along with an improved Deep Belief Network (E-DBN) to compute the visual information from the given images. Further, for the classification task, the approach used numerous classifiers, with the highest value attained by the SVM classifier with an accuracy of 99.161%. Alsheikhy et al. (59) used the VGG-19 model with long short-term memory networks (LSTMs) to execute the classification task of lung cancer nodules. This technique has stated an accuracy value of 99.42%. The results depicted in Table 6 indicate that the Residual-SwishNet approach has attained the highest results in terms of all evaluation parameters used in the assessment compared to all the modern approaches. We have reported an accuracy score of 99.60% with a performance gain of 2.31%. Further, in terms of precision, the proposed model has reported a performance gain of 3.98%, which is 3.21% and 3.18% for the recall and F1-Score, respectively.
The main cause of attaining effective scores in comparison to the latest approaches is because of the robust information learning capability of the suggested solution. The work in (27) improved features recalibration but lacked advanced activation mechanisms to compute complex lung patterns. Further, the work in (60) lacks to handle the distorted samples. Similarly, a CNN with attention in (49) enhanced focus on key regions but was limited by the shallow feature extraction backbone. Further, the approach in (56) followed a multi-stage pipeline combining which increased complexity and risked cumulative errors. Earlier models, like improved AlexNet with SVM (57) and Gabor filters combined with Deep Belief Networks (E-DBN) and SVM (58), relied on outdated or manually crafted features, restricting their depth and adaptability. Furthermore, the method in (59) computed dense features; however, it suffered from the model overfitting issue. In contrast, our proposed Residual-SwishNet overcomes these limitations through an optimized ResNet50 backbone integrated with the Swish activation function, enhancing nonlinear learning and preserving critical negative activations. Additionally, we introduced three dense layers for richer feature abstraction, creating a streamlined, end-to-end framework that efficiently learns from lung CT images and outperforms prior methods in results and robustness on the LUNA16 dataset.
Next, we have carried out the comparative analysis of the Residual-SwishNet for the IQ-OTH/NCCD dataset against the latest approaches (61–65), and the obtained values are provided in Table 7.
Kumar et al. (61) suggested a DL approach for lung cancer classification, where VGG19 was used to compute deep features, while for the classification, the Vision Transformer (ViT) was applied. This work attained an accuracy of 99.06%. The work in (62) presented a hybrid approach employing various DL methods like GoogLeNet, EfficientNet, DarkNet19, and ResNet18 to perform the diagnosis of lung tumors, and stated an accuracy of 99%. Venkatraman et al. (63) also designed a DL framework for lung nodules classification, in which VGG16 was used to compute deep features, while for the classification task, the SVM approach was used. The work reported an accuracy value of 89.36%. Further, the work in (64) employed an improved GoogLeNet approach with Adaptive Layers named GoogLeNet-AL for recognizing lung cancer nodules. The technique has acquired a categorization score of 97.32%. Gupta et al. (65) proposed a DL method for lung cancer classification that presented an enhanced U-Net framework in which a conventional U-Net model was used for multi-scale features computation along with the Differentiable Architecture Search. The work achieved an accuracy of 98.82%. In comparison, the suggested framework again attained the highest values for all measures over the IQ-OTH/NCCD dataset in comparison to all approaches. The improved feature engineering capability, along with the high recall rate of the proposed approach, assists the model in attaining robust results. On the IQ-OTH/NCCD dataset, prior studies revealed several architectural and methodological limitations. For instance, the work in (61) combined VGG19 with ViT, but the shallow feature extraction of VGG19 and the data-hungry nature of ViT limited performance on smaller medical datasets. Similarly, ensemble models in (62) introduced unnecessary computational overhead without guaranteeing significant performance gains. Other works, like those discussed in (63, 64), were constrained by outdated or shallow backbones that struggled to extract the complicated visual patterns of lung cancer. Segmentation-based models like those discussed in (65) added additional pipeline complexity, which was not optimal for direct classification tasks. In contrast, our Residual-SwishNet effectively integrates a deep ResNet50 backbone with Swish activation and added dense layers, enabling rich feature learning without excessive complexity. Further, the inclusion of cross-entropy loss in the final classification layer helps to tackle the class-imbalance problem. The overall architectural description of the DeepLungNe assists in computing fine-grained patterns by minimizing training inefficiencies and leads to superior classification results across key performance metrics.
4.7 Cross-dataset evaluation
To evaluate the generalization capability of the proposed model, we performed cross-dataset experiments using the benign and malignant classes from two different datasets: LUNA-16 and IQ-OTH/NCCD, and attained results are provided in Table 8. In the first setting, the model was trained on the LUNA-16 dataset and tested on IQ-OTH/NCCD, and attained an accuracy of 65.41%. In the second setting, the model was trained on IQ-OTH/NCCD and tested on LUNA-16, resulting in an accuracy of 59.93%.
These results show that the model retains a reasonable level of recognition ability even when tested on entirely unseen datasets, which shows that the learned features by the proposed approach possess a degree of transferability. The better performance when trained on LUNA-16 is attributed to its larger size and greater variability, which allow the Residual-SwishNet to learn richer and more generalizable representations. However, the drop in accuracy in both cases highlights the challenges posed by differences in image acquisition protocols, resolution, and noise characteristics between datasets. Although the cross-dataset results are promising, there is still room for improvement. Incorporating advanced domain adaptation methods, more robust data augmentation strategies, or transfer learning fine-tuning could help bridge the performance gap and make the model more resilient to domain shifts, ultimately enhancing its real-world applicability.
5 Conclusion
This study has introduced a DL approach named Residual-SwishNet for lung cancer classification. Specifically, we altered the ResNet50 framework by integrating the Swish activation function, additional dense layers, and a Softmax output with Cross-Entropy Loss to enhance learning capacity and classification precision. The work is evaluated on two standard datasets, named the LUNA16 and IQ-OTH/NCCD datasets, and achieved accuracy scores of 99.60% and 99.11%, outperforming existing state-of-the-art methods. These results highlight the competency of our approach to accurately differentiate between benign and malignant lung nodules by offering potential support in early diagnosis. However, one limitation of our approach is the increased computational complexity and training time introduced by model modifications, which can limit deployment on resource-constrained devices or real-time applications. The proposed Residual-SwishNet carries strong clinical relevance, as accurate early differentiation between benign and malignant lung nodules can support radiologists and reduce diagnostic delays. However, the study has certain limitations, including the use of only two publicly available datasets, which may introduce dataset-specific biases and limit generalizability. External validation on multi-center clinical data and more diverse patient populations is still required. Additionally, although the cross-dataset evaluation shows promising transferability, the performance gap highlights the need for further robustness improvements. Future research will focus on integrating domain adaptation techniques, incorporating 3D volumetric information, and validating the model in real clinical workflows to enhance reliability and practical adoption.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/datasets/avc0706/luna16, https://www.kaggle.com/datasets/hamdallak/the-iqothnccd-lung-cancer-dataset.
Ethics statement
Ethical approval was not required for the studies on humans in accordance with the local legislation and institutional requirements because only commercially available established cell lines were used.
Author contributions
MN: Conceptualization, Data curation, Formal analysis, Methodology, Software, Visualization, Writing – original draft, Writing – review & editing. AJ: Conceptualization, Formal analysis, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing. AS: Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Validation, Visualization, Writing – original draft, Writing – review & editing.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2601).
Acknowledgments
The authors extend their appreciation to the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) for funding this work through (grant number IMSIU-DDRSP2601).
Conflict of interest
The authors declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Wani NA, Kumar R, and Bedi J. DeepXplainer: An interpretable deep learning based approach for lung cancer detection using explainable artificial intelligence. Comput Methods Programs Biomed. (2024) 243:107879. doi: 10.1016/j.cmpb.2023.107879
2. Gomez DR and Liao Z. Non-small cell lung cancer (NSCLC) and small cell lung cancer (SCLC). In: Target Volume Delineation and Field Setup: A Practical Guide for Conformal and Intensity-Modulated Radiation Therapy. Berlin, Heidelberg: Springer (2012). p. 87–103.
3. Leiter A, Veluswamy RR, and Wisnivesky JP. The global burden of lung cancer: current status and future trends. Nat Rev Clin Oncol. (2023) 20:624–39. doi: 10.1038/s41571-023-00798-3
4. Siegel RL, Miller KD, Wagle NS, and Jemal A. Cancer statistics, 2023. CA: Cancer J Clin. (2023) 73:17–48. doi: 10.3322/caac.21763
5. Padinharayil H, Varghese J, John MC, Rajanikant GK, Wilson CM, Al-Yozbaki M, et al. Non-small cell lung carcinoma (NSCLC): Implications on molecular pathology and advances in early diagnostics and therapeutics. Genes Dis. (2023) 10:960–89. doi: 10.1016/j.gendis.2022.07.023
6. Vadala R, Pattnaik B, Bangaru S, Rai D, Tak J, Kashyap S, et al. A review on electronic nose for diagnosis and monitoring treatment response in lung cancer. J Breath Res. (2023) 17. doi: 10.1088/1752-7163/acb791
7. Lanjewar MG, Panchbhai KG, and Charanarur P. Lung cancer detection from CT scans using modified DenseNet with feature selection methods and ML classifiers. Expert Syst Appl. (2023) 224:119961. doi: 10.1016/j.eswa.2023.119961
8. Makaju S, Prasad P, Alsadoon A, Singh A, and Elchouemi A. Lung cancer detection using CT scan images. Proc Comput Sci. (2018) 125:107–14. doi: 10.1016/j.procs.2017.12.016
9. Pastorino U, Bellomi M, Landoni C, De Fiori E, Arnaldi P, Picchio M, et al. Early lung-cancer detection with spiral CT and positron emission tomography in heavy smokers: 2-year results. Lancet. (2003) 362:593–7. doi: 10.1016/S0140-6736(03)14188-8
10. Pan X, Dvortsin E, Baldwin DR, Groen HJM, Ramaker D, Ryan J, et al. Cost-effectiveness of volume computed tomography in lung cancer screening: a cohort simulation based on nelson study outcomes. J Med Econ. (2024) 27:27–38. doi: 10.1080/13696998.2023.2288739
11. Svoboda E. Artificial intelligence is improving the detection of lung cancer. Nature. (2020) 587:S20–0. doi: 10.1038/d41586-020-03157-9
12. Xu K, Khan MS, Li TZ, Gao R, Terry JG, Huo Y, et al. AI body composition in lung cancer screening: added value beyond lung cancer detection. Radiology. (2023) 308:e222937. doi: 10.1148/radiol.222937
13. Sathyakumar K, Munoz M, Singh J, Hussain N, Babu BA, and Babu B. Automated lung cancer detection using artificial intelligence (AI) deep convolutional neural networks: A narrative literature review. Cureus. (2020) 12. doi: 10.7759/cureus.10017
14. Joshua ESN, Chakkravarthy M, and Bhattacharyya D. An extensive review on lung cancer detection using machine learning techniques: A systematic study. Rev d’Intelligence Artificielle. (2020) 34:351–19. doi: 10.18280/ria.340314
15. Chassagnon G, De Margerie-Mellon C, Vakalopoulou M, Marini R, Hoang-Thi TN, Revel MP, et al. Artificial intelligence in lung cancer: current applications and perspectives. Japanese J Radiol. (2023) 41:235–44. doi: 10.1007/s11604-022-01359-x
16. de Margerie-Mellon C and Chassagnon G. Artificial intelligence: A critical review of applications for lung nodule and lung cancer. Diagn Interventional Imaging. (2023) 104:11–7. doi: 10.1016/j.diii.2022.11.007
17. Hosseini SH, Monsefi R, and Shadroo S. Deep learning applications for lung cancer diagnosis: a systematic review. Multimedia Tools Appl. (2023) 83:1–31. doi: 10.1007/s11042-023-16046-w
18. Huang M, Zou J, Zhang Y, Bhatti UA, and Chen J. Efficient click-based interactive segmentation for medical image with improved Plain-ViT. IEEE J Biomed Health Inf. (2024) 29:8904–16. doi: 10.1109/JBHI.2024.3392893
19. Huang M, Zhang XS, Bhatti UA, Wu Y, Zhang Y, and Ghadi YY. An interpretable approach using hybrid graph networks and explainable AI for intelligent diagnosis recommendations in chronic disease care. Biomed Signal Process Control. (2024) 91:105913. doi: 10.1016/j.bspc.2023.105913
20. Nizamani AH, Chen Z, Nizamani AA, and Bhatti UA. Advance brain tumor segmentation using feature fusion methods with deep U-Net model with CNN for MRI data. J King Saud University-Computer Inf Sci. (2023) 35:101793. doi: 10.1016/j.jksuci.2023.101793
21. Cellina M, Cacioppa LM, Ce M, Chiarpenello V, Costa M, Vincenzo Z, et al. Artificial intelligence in lung cancer screening: the future is now. Cancers. (2023) 15:4344. doi: 10.3390/cancers15174344
22. Bhatti UA, Huang M, Neira-Molina H, Marjan S, Baryalai M, Tang H, et al. MFFCG–Multi feature fusion for hyperspectral image classification using graph attention network. Expert Syst Appl. (2023) 229:120496. doi: 10.1016/j.eswa.2023.120496
23. Bhatti UA, Tang H, Wu G, Marjan S, and Hussain A. Deep learning with graph convolutional networks: An overview and latest applications in computational intelligence. Int J Intelligent Syst. (2023) 2023:8342104. doi: 10.1155/2023/8342104
24. Raza R, Zulfiqar F, Khan MO, Arif M, Alvi A, Iftikhar MA, et al. Lung-EffNet: Lung cancer classification using EfficientNet from CT-scan images. Eng Appl Artif Intell. (2023) 126:106902. doi: 10.1016/j.engappai.2023.106902
25. Nahiduzzaman M, Abdulrazak LF, Ayari MA, Khandakar A, and Islam SR. A novel framework for lung cancer classification using lightweight convolutional neural networks and ridge extreme learning machine model with SHapley Additive exPlanations (SHAP). Expert Syst Appl. (2024) 248:123392. doi: 10.1016/j.eswa.2024.123392
26. Gulsoy T and Kablan EB. FocalNeXt: A ConvNeXt augmented FocalNet architecture for lung cancer classification from CT-scan images. Expert Syst Appl. (2025) 261:125553. doi: 10.1016/j.eswa.2024.125553
27. Priya A and Bharathi PS. SE-resNeXt-50-CNN: A deep learning model for lung cancer classification. Appl Soft Computing. (2025) 171:112696. doi: 10.1016/j.asoc.2025.112696
28. Mohandass G, Krishnan GH, Selvaraj D, and Sridhathan C. Lung cancer classification using optimized attention-based convolutional neural network with DenseNet-201 transfer learning model on CT image. Biomed Signal Process Control. (2024) 95:106330. doi: 10.1016/j.bspc.2024.106330
29. Murthy NN and Thippeswamy K. TPOT with SVM hybrid machine learning model for lung cancer classification using CT image. iomedical Signal Process Control. (2025) 104:107465. doi: 10.1016/j.bspc.2024.107465
30. Mothkur R and Veerappa B. Classification of lung cancer using lightweight deep neural networks. Proc Comput Sci. (2023) 218:1869–77. doi: 10.1016/j.procs.2023.01.164
31. Sangeetha S, Mathivanan SK, Karthikeyan P, Rajadurai H, Shivahare BD, Mallik S, et al. An enhanced multimodal fusion deep learning neural network for lung cancer classification. Syst Soft Computing. (2024) 6:200068. doi: 10.1016/j.sasc.2023.200068
32. Mohamed TI, Oyelade ON, and Ezugwu AE. Automatic detection and classification of lung cancer CT scans based on deep learning and ebola optimization search algorithm. PloS One. (2023) 18:e0285796. doi: 10.1371/journal.pone.0285796
33. Hossain MS, Basak N, Mollah MA, Nahiduzzaman M, Ahsan M, and Haider J. Ensemble-based multiclass lung cancer classification using hybrid CNN-SVD feature extraction and selection method. PloS One. (2025) 20:e0318219. doi: 10.1371/journal.pone.0318219
34. Naseer I, Akram S, Masood T, Rashid M, and Jaffar A. Lung cancer classification using modified u-net based lobe segmentation and nodule detection. IEEE Access. (2023) 11:60279–91. doi: 10.1109/ACCESS.2023.3285821
35. Uddin AH, Chen Y-L, Akter MR, Ku CS, Yang J, and Por LY. Colon and lung cancer classification from multi-modal images using resilient and efficient neural network architectures. Heliyon. (2024) 10. doi: 10.1016/j.heliyon.2024.e30625
36. Pandit BR, Alsadoon A, Prasad PWC, Al Aloussi S, Rashid TA, Alsadoon OH, et al. Deep learning neural network for lung cancer classification: enhanced optimization function. Multimedia Tools Appl. (2023) 82:6605–24. doi: 10.1007/s11042-022-13566-9
37. Alsallal M, Ahmed HH, Kareem RA, Yadav A, Ganesan S, Shankhyan A, et al. Enhanced lung cancer subtype classification using attention-integrated DeepCNN and radiomic features from CT images: a focus on feature reproducibility. Discover Oncol. (2025) 16:336. doi: 10.1007/s12672-025-02115-z
38. He K, Zhang X, Ren S, and Sun J. (2016). Deep residual learning for image recognition, in: Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA: IEEE. pp. 770–8.
39. Nawaz M, Javed A, and Irtaza A. ResNet-Swish-Dense54: a deep learning approach for deepfakes detection. Visual Comput. (2022) 39:1–22. doi: 10.1007/s00371-022-02732-7
40. Patwardhan N, Ingalhalikar M, and Walambe R. ARiA: utilizing richard’s curve for controlling the non-monotonicity of the activation function in deep neural nets. arXiv preprint arXiv:.08878. (2018). doi: 10.48550/arXiv.1805.08878
41. Su C and Wang W. Concrete cracks detection using convolutional neuralNetwork based on transfer learning. Math Problems Eng. (2020) 2020. doi: 10.1155/2020/7240129
42. Neal Joshua ES, Bhattacharyya D, Chakkravarthy M, and Byun Y-C. 3D CNN with visual insights for early detection of lung cancer using gradient-weighted class activation. J Healthcare Eng. (2021) 2021:1–11. doi: 10.1155/2021/6695518
43. Al-Yasriy HF, Al-Husieny MS, Mohsen FY, Khalil EA, and Hassan ZS. (2020). Diagnosis of lung cancer based on CT scans using CNN, in: IOP Conference Series: Materials Science and Engineering, Vol. 928. p. 022035. IOP Publishing.
44. Albattah W, Javed A, Nawaz M, and Albahli S. Artificial intelligence-based drone system for multiclass plant disease detection using an improved efficient convolutional neural network. Front Plant Sci. (2022) 13:808380. doi: 10.3389/fpls.2022.808380
45. Adedoja AO, Owolawi PA, Mapayi T, and Tu C. Intelligent mobile plant disease diagnostic system using NASNet-mobile deep learning. IAENG Int J Comput Sci. (2022) 49:216–31.
46. Chhabra M and Kumar R. A smart healthcare system based on classifier DenseNet 121 model to detect multiple diseases. In: Mobile Radio Communications and 5G Networks: Proceedings of Second MRCN 2021. Singapore: Springer (2022). p. 297–312.
47. Sandler M, Howard A, Zhu M, Zhmoginov A, and Chen L-C. (2018). Mobilenetv2: Inverted residuals and linear bottlenecks, in: Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA: IEEE. pp. 4510–20.
48. Chen H-Y and Su C-Y. (2018). An enhanced hybrid MobileNet, in: 2018 9th International Conference on Awareness Science and Technology (iCAST), . pp. 308–12. Fukuoka, Japan: IEEE.
49. UrRehman Z, Qiang Y, Wang L, Shi Y, Yang Q, Khattak SU, et al. Effective lung nodule detection using deep CNN with dual attention mechanisms. Sci Rep. (2024) 14:3934. doi: 10.1038/s41598-024-51833-x
50. Xia X, Xu C, and Nan B. (2017). Inception-v3 for flower classification, in: 2017 2nd international conference on image, vision and computing (ICIVC). pp. 783–7. Chengdu: IEEE.
51. Koonce B. ResNet 50. In: Convolutional neural networks with swift for tensorflow: image recognition and dataset categorization. Berkeley, CA, USA: Springer (2021). p. 63–72.
52. Zhang L, Li H, Zhu R, and Du P. An infrared and visible image fusion algorithm based on ResNet-152. Multimedia Tools Appl. (2022) 81:9277–87. doi: 10.1007/s11042-021-11549-w
53. Panigrahi U, Sahoo PK, Panda MK, and Panda G. A ResNet-101 deep learning framework induced transfer learning strategy for moving object detection. Image Vision Computing. (2024) 146:105021. doi: 10.1016/j.imavis.2024.105021
54. Chaddad A, Peng J, and Wu Y. Classification based deep learning models for lung cancer and disease using medical images. arXiv preprint arXiv:.01279. (2025). doi: 10.1109/TRPMS.2025.3584031
55. Roy S, Singh S, Sahu P, Suresh A, and Das D. MSAD-net: multiscale and spatial attention-based dense network for lung cancer classification. arXiv preprint arXiv:.14626. (2025). doi: 10.48550/arXiv.2504.14626
56. Thangavel C and Palanichamy J. Effective deep learning approach for segmentation of pulmonary cancer in thoracic CT image. Biomed Signal Process Control. (2024) 89:105804. doi: 10.1016/j.bspc.2023.105804
57. Naseer I, Masood T, Akram S, Jaffar A, Rashid M, and Iqbal MA. Lung cancer detection using modified alexNet architecture and support vector machine. Comp Mat Continua (2023) 74:2039–54. doi: 10.32604/cmc.2023.032927
58. Siddiqui EA, Chaurasia V, and Shandilya M. Detection and classification of lung cancer computed tomography images using a novel improved deep belief network with Gabor filters. Chemometrics Intelligent Lab Syst. (2023) 235:104763. doi: 10.1016/j.chemolab.2023.104763
59. Alsheikhy AA, Said Y, Shawly T, Alzahrani AK, and Lahza H. A CAD system for lung cancer detection using hybrid deep learning techniques. Diagnostics. (2023) 13:1174. doi: 10.3390/diagnostics13061174
60. Faizi MK, Qiang Y, Wei Y, Qiao Y, Zhao J, Aftab R, et al. Deep learning-based lung cancer classification of CT images. BMC Cancer. (2025) 25:1056. doi: 10.1186/s12885-025-14320-8
61. Kumar A, Ansari MK, and Singh KK. A combined approach of vision transformer and transfer learning based model for accurate lung cancer classification. SN Comput Sci. (2025) 6:1–11. doi: 10.1007/s42979-025-04050-w
62. Güraksın GE and Kayadibi I. A hybrid LECNN architecture: a computer-assisted early diagnosis system for lung cancer using CT images. Int J Comput Intell Syst. (2025) 18:35. doi: 10.1007/s44196-025-00761-3
63. Venkatraman K and Reddy SNPS. (2024). Augmenting clinical decisions with deep learning lung cancer image abnormality segmentation, in: 2024 14th International Conference on Cloud Computing, Data Science & Engineering (Confluence). pp. 674–8. Noida, India: IEEE.
64. Ma L, Wu H, and Samundeeswari P. Googlenet-al: A fully automated adaptive model for lung cancer detection. Pattern Recognition. (2024) 155:110657. doi: 10.1016/j.patcog.2024.110657
Keywords: classification, deep learning, lung cancer, ResNet50, Swish
Citation: Nawaz M, Javed A and Saudagar AKJ (2026) Residual-SwishNet: a deep learning-based approach for reliable lung cancer classification. Front. Oncol. 15:1729021. doi: 10.3389/fonc.2025.1729021
Received: 01 November 2025; Accepted: 18 December 2025; Revised: 06 December 2025;
Published: 26 January 2026.
Edited by:
Shagun Sharma, VIT Bhopal University, IndiaReviewed by:
Vivek Bhardwaj, Amity University, Mohali, IndiaPriyanka Datta, Laxmi Institute of Technology, India
Copyright © 2026 Nawaz, Javed and Saudagar. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Abdul Khader Jilani Saudagar, YWtzYXVkYWdhckBpbWFtdS5lZHUuc2E=