Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell., 12 August 2025

Sec. AI in Food, Agriculture and Water

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1640549

Enhanced plant disease classification with attention-based convolutional neural network using squeeze and excitation mechanism

S. KarthikeyanS. Karthikeyan1R. CharanR. Charan1Sathiya Narayanan
Sathiya Narayanan1*L. Jani AnbarasiL. Jani Anbarasi2
  • 1School of Electronics Engineering, Vellore Institute of Technology, Chennai, India
  • 2School of Computer Science Engineering, Vellore Institute of Technology, Chennai, India

Introduction: Technology is becoming essential in agriculture, especially with the growth of smart devices and edge computing. These tools help boost productivity by automating tasks and allowing real-time analysis on devices with limited memory and resources. However, many current models struggle with accuracy, size, and speed particularly when handling multi-label classification problems.

Methods: This paper proposes a Convolutional Neural Network with Squeeze and Excitation Enabled Identity Blocks (CNN-SEEIB), a hybrid CNN-based deep learning architecture for multi-label classification of plant diseases. CNN-SEEIB incorporates an attention mechanism in its identity blocks to leverage the visual attention that enhances the classification performance and computational efficiency. PlantVillage dataset containing 38 classes of diseased crop leaves alongside healthy leaves, totaling 54,305 images, is utilized for experimentation.

Results: CNN-SEEIB achieved a classification accuracy of 99.79%, precision of 0.9970, recall of 0.9972, and an F1 score of 0.9971. In addition, the model attained an inference time of 64 milliseconds per image, making it suitable for real-time deployment. The performance of CNNSEEIB is benchmarked against the state-of-the-art deep learning architectures, and resource utilization metrics such as CPU/GPU usage and power consumption are also reported, highlighting the model’s efficiency.

Discussion: The proposed architecture is also validated on a potato leaf disease dataset of 4,062 images from Central Punjab, Pakistan, achieving a 97.77% accuracy in classifying Healthy, Early Blight, and Late Blight classes.

1 Introduction

The importance of agriculture in everyday life cannot be minimized, as it offers an overabundance of direct and indirect benefits. Primarily, agriculture provides food for living beings as a direct benefit. Beyond this fundamental function, agriculture yields a range of indirect benefits that are equally vital. Moreover, agriculture significantly contributes to the Gross Domestic Product (GDP) showcasing its multifaceted impact on economy and the society (U.S. Department of Agriculture, 2020). Despite its positive contributions, there are challenges hindering the growth and production within this sector. One major issue is the occurrence of diseases in plants, which can impede growth and production. When a plant becomes diseased, not only does it suffer, but nearby plants are also at risk of infection, leading to a decline in overall production. Consequently, this affects both the GDP and the direct and indirect benefits associated with agriculture. Based on statistics, global crop loss is estimated at $220 billion (United States Dollars) annually, with plant disease accounting for 14.1% of this total (National Institute of Food and Agriculture, 2022). Biotic factors such as oomycetes, fungi, viruses, bacteria, nematodes, and viroids, along with abiotic elements like environmental factors contribute to this loss. Various rust diseases of wheat accounts to a revenue loss of approximately £60,000,000 worldwide, and India’s share is approximately £4,000,000 (Bayer Crop Science, n.d.). The Jowar crop affected by the Smut disease (SlideShare, n.d.), due to the fungus called Sphacelotheca Sorghi, resulted in a revenue loss of approximately £2,000,000 in the Mumbai region alone. Plant diseases affect crop quality, shelf life, and nutritional value, ultimately reducing yields and marketability. In addition, the impact of climate change and urbanization intensifies pressure on global food production systems. Earlier, manual techniques involving the identification of diseased areas and discerning the type of disease were carried out with the aid of experts and utilized a large amount of labour, time, and cost for the process. Manual disease classification is a slow and time-consuming process, often lacking accuracy, especially when dealing with numerous diseased plants. Several disease identification algorithms based on Artificial Intelligence (AI), Machine Learning (ML), and Deep Learning (DL) were introduced for automating the classification of diseased and healthy plants.

Automated plant disease classification has significantly improved the accuracy and efficiency of identifying crop issues. These advancements are helping farmers to make better decisions about plant health and management. With the integration of cutting-edge techniques, farming practices are made more sustainable, resilient, and resource-efficient. This paper proposes a deep learning architecture called Convolutional Neural Network with Squeeze and Excitation Enabled Identity Blocks (CNN-SEEIB), which integrates Squeeze-and-Excitation (SE) attention mechanisms within identity blocks for efficient plant disease classification. This light-weight model is ideal for edge devices with limited resources. It enhances feature learning by focusing on the most important information, resulting in improved classification accuracy. Designed for real-time inference, it ensures quick and efficient predictions.

The main contributions of the proposed CNN-SEEIB are as follows.

• The proposed CNN-SEEIB is a customized lightweight backbone with fewer parameters integrated with SE attention to enhance feature representation, enabling efficient, accurate, and real-time multi-label plant disease classification.

• A comparative study with 8 different pre-trained models demonstrated that CNN-SEEIB outperforms several state-of-the-art models. The accuracy of CNN-SEEIB was also compared with several state-of-the-art approaches for plant disease classification.

• The proposed model’s utilization of system resources, including CPU and GPU usage and power consumption, is evaluated and analysed as percentage metrics. The analysis of CNN-SEEIB in this study focuses on architecture optimization, which involves refining the traditional network structure and components to enhance overall performance.

With the growing demand for real-time and efficient solutions, the proposed CNN-SEEIB addresses the shortcomings of existing DL models, mainly in multi-label classification. The CNN-SEEIB architecture attained an optimized and efficient model for accurate leaf disease prediction. The proposed model resulted in better accuracy by leveraging visual attention to prioritize key features, without significantly increasing model complexity or size. Integrating SE blocks enhances feature representation and allows the model to adaptively highlight important features and suppress less significant ones, improving both accuracy and efficiency. The optimization of computational requirements, particularly for edge devices, is important for real-world agricultural applications where resource constraints are significant. Deploying the CNN-SEEIB model on portable edge devices such as drones or handheld diagnostic tools can enable real-time plant disease detection directly in the field.

The paper is structured as follows: Section 2 discusses the existing studies related to identifying and classifying plant diseases. Section 3 explains the proposed CNN-SEEIB model along with its computational complexity and highlights. Section 4 presents the results and discussion that details the dataset utilized, experimentation, and analysis of the obtained results. Section 5 concludes with the future scope.

2 Literature survey

Over the years, visual and image processing has seen a significant improvement in the number of algorithms in various domains, coinciding with the boom in the complexity of the problems addressed. In the realm of agriculture, for identifying and classifying plant diseases, a wide variety of techniques and methodologies have been proposed. With simple image processing and ML techniques, the field has progressed to embrace the power of AI and advanced DL approaches, offering comprehensive end-to-end solutions for addressing the challenges posed.

2.1 Image processing and machine learning based approaches

Vamsidhar et al. (2019) analysed co-occurrence feature computation and classifiers like K-means, Support Vector Machines (SVM), and MultiLayer Perceptron (MLP) for leaf disease classification. Leaf images were masked by filtering out green pixels, leaving behind the infected regions. Hybrid clustering, combining K-means and hierarchical clustering, segmenting the infected areas, further improved by a genetic algorithm for disease categorization. Co-occurrence features were selected for unsupervised ML techniques like K-means, while supervised methods like Naïve Bayes, SVM, and MLP were directly used without explicit feature selection, achieving an accuracy of 95.87% with SVM. Similarly, to identify disease in tomato plant leaves, Xian and Ngadiran (2021) proposed an Extreme Learning Machine (ELM) classifier with a single-layer feed-forward neural network. Harlick textures for feature extraction were used to the pre-processed image, and converted from Red-Green-Blue (RGB) to hue-saturation-value color space, resulting in an accuracy of 84.94%.

Singla et al. (2023) used machine learning approach using Neural Networks, SVM, and Naïve Bayes to classify plant leaves as healthy or infected. This method involved extracting features with a CNN, labelling the classes, and assigning probabilities through logistic regression. These classifiers were then used to determine the leaf categories, with the Neural Network classifier achieving an accuracy of 94% on the dataset. Ramesh et al. (2018) used a Random Forest (RF) classifier to distinguish between healthy and diseased papaya leaves. This method extracted feature descriptors using Histogram of Oriented Gradients, captured leaf edges with Hue moments, analysed texture using Haralick features, and represented pixel distribution with Color Histograms. The RF classifier trained on these features achieved an accuracy of 71%, outperforming other classifiers. Sabrol and Satish (2016) classified six categories of tomato leaves and stems, one healthy and five diseased by extracting shape, color, and texture features after image segmentation. Classification tree based on these features were developed, which achieved an overall testing accuracy of about 97.3%.

Thomkaew and Intakosum (2023) proposed a novel method for classifying plant leaves by combining CNN with the Scale Invariant Feature Transform (SIFT) algorithm. Leaf images were converted to binary followed by morphological transformations, like opening and closing that removed noise through erosion and dilation. Canny Edge detection was performed to identify edges in this processed images, and key-point features were extracted using SIFT. These processed images were then processed using a CNN for feature extraction, followed by classification using a Random Forest classifier. The approach achieved an accuracy of 95.62%.

2.2 Convolutional neural network based approaches

Deep learning techniques like CNNs has transformed computer vision, expanding its applications from classification to tasks like object detection, segmentation, and pose estimation, while increasing the algorithm performance across various fields (Alzubaidi et al., 2021). Specifically, CNNs have shown great success in identifying and classifying plant diseases (Tuğrul et al., 2022; Saleem et al., 2020; Maeda-Gutiérrez et al., 2020), leading to models that deliver impressive accuracy and efficient resource use. Batchuluun et al. (2022) introduced Periodic Implicit CNN (PI-CNN), a 30-layer CNN combining convolutional and residual blocks, which performed well on smaller datasets. They also developed PI-GAN, a Generative Adversarial Network (GAN) for augmenting dataset samples. Datasets like PlantVillage, PlantDoc, Fruits360, and Plants were evaluated based on the metrics like true positive rate, positive predictive value, F1-score, and accuracy.

Majji and Kumaravelan (2021) developed a CNN to classify 38 disease categories from the PlantVillage dataset, using only 16% of the total samples. The model, trained with a 90 to 10% split, batch size of 32, and 0.5 dropout over 200 epochs, attained an accuracy of 99.89%. Peyal et al. (2023) developed a lightweight 11-layer 2D-CNN for smartphone-assisted disease diagnosis, incorporated as an Android app called Plant Disease Classifier. This model classified 14 classes of tomato and cotton diseases with an average accuracy of 97.36% and an AUC of 99.9%. It also outperformed transfer learning models such as Inception V3, VGG16, VGG19, and MobileNet V1 and V2, while achieving a fast inference speed of 4.84 ms.

Imanulloh et al. (2023) designed a simple 12-layer CNN to classify plant diseases using the augmented PlantVillage dataset. Their model achieved accuracy, precision, recall, and F1-score of 97, 98, 97, and 97% respectively, outperforming other ImageNet models like VGG16, MobileNetV2, DenseNet121, ResNet50, and InceptionV3. Pandian et al. (2022) proposed a 14-layer deep CNN to classify 58 leaf disease classes across 16 plant species. They combined five datasets and applied augmentation techniques including Basic Image Manipulation (BIM), Neural Style Transfer (NST), and GANs to address class imbalance. Trained on multiple GPUs, this model achieved a classification accuracy of 99.97%, with weighted average precision, recall, and F1 scores of approximately 99.8%.

Naik et al. (2022) performed comparative analysis of twelve pre-trained deep learning architectures, such as VGG19 and DarkNet53, to classify five major chilli leaf diseases using a custom image dataset. VGG19 attained the highest accuracy of 83.54% on non-augmented data, whereas DarkNet53 outperformed others with an accuracy of 98.82% on augmented data. A custom squeeze-and-excitation-based CNN (SECNN) further improved performance, reaching 98.63% accuracy without augmentation and 99.12% with it. Furthermore, the SECNN model was evaluated on a plant leaf disease class from the PlantVillage dataset, resulting a robust generalization with an overall accuracy of 99.28%. Bhuyan et al. (2023) proposed SE_SPnet, a stacked parallel convolutional neural network integrated with a squeeze-and-excitation (SE) block for classifying rice leaf diseases. The architecture integrated multiple convolutional layers with varying kernel sizes to capture both local and global features, while the SE block enhanced relevant feature extraction. The model attained the accuracy. Sensitivity, specificity, precision, recall and F1-score as 99.2, 98.2, 98.5, 98.4, 98.2 and 98.5%, respectively, using SGD with momentum and 0.01 learning rate.

Naresh et al. (2024) proposed an optimized Squeeze-and-Excitation Densely Connected Convolutional Neural Network (SEDCNN) for early detection of chili leaf diseases. Among various SE block configurations, the integration of the standard SE block with SEDCNN achieved the best performance with 97% accuracy. With data augmentation, the model’s accuracy further improved to 98.86%, outperforming conventional CNNs, ResNet variants, and multiple transfer learning models. Chen et al. (2021) proposed SE-MobileNet, a hybrid model combining the lightweight MobileNet architecture with squeeze-and-excitation (SE) blocks to enhance plant disease identification. A two-phase transfer learning approach was used: the first phase involved training extended layers, followed by fine-tuning the entire model in the second phase. The SE-MobileNet model achieved a high accuracy of 99.78% on a public dataset with clear backgrounds and 99.33% under complex conditions with multiple disease classes.

Ashurov et al. (2025) proposed depthwise CNN model integrated with squeeze-and-excitation blocks, residual skip connections for accurate plant disease detection. The model attained an accuracy of 98% and an F1-score of 98.2%, resulting as a reliable tool for practical agricultural applications. Assaduzzaman et al. (2025) presented XSE-TomatoNet, an improved EfficientNetB0-based model enhanced with Squeeze-and-Excitation (SE) blocks and multi-scale feature fusion for precise tomato leaf disease classification. The model achieved 99.11% as accuracy and 99% precision and recall by 10-fold cross-validation.

Singh and Singh (2025) enhanced the ResNet50V2 model by integrating a Squeeze-and-Excitation (SE) block to improve accuracy in rice disease identification. Using the Kaggle rice leaf diseases dataset, the SE-enhanced model achieved a testing accuracy of 93.33%, outperforming the original ResNet50V2 which scored 83.61%. With a slight increase in parameters, the performance gain demonstrates the effectiveness of channel-wise attention through SE blocks.

The vine plant is economically important not only for grapes but also for products like wine, molasses, and culinary-use grape leaves. Diseases affecting grape leaves reduce yield and make the leaves unusable, leading to significant financial losses. Unal (2025) focused on classifying common grape leaf diseases such as scab and downy mildew alongside healthy leaves using deep learning models. The pre-trained networks performance was enhanced by integrating Convolutional Block Attention Module (CBAM) and Squeeze-and-Excitation (SE) blocks. These improvements increased the classification accuracy from 92.73 to 96.36%, demonstrating the effectiveness of attention-based enhancements. Nikhileswar et al. (2024) proposed custom CNN model named PlantLDNet optimized to distinguish diseased and healthy leaves while mitigating vanishing gradient problem. The model’s performance was compared with previous methods using metrics such as precision, recall, F1-score, ROC curves, and AUC. PlantLDNet attained F1-scores as 93% for Early Bright, 98% for Healthy, and 92% for Late Bright conditions.

2.3 Transfer learning based approaches

Optimizing pre-trained models (Pan and Yang, 2010) allows fine-tuning to fit specific tasks thus enhancing the performance through leveraging the strengths of the original models while saving training time (Balafas et al., 2023). Panchal et al. (2021) analysed transfer learning models like VGG16, ResNet50, and InceptionV3 for classifying leaf images into four infection stages (Healthy, Early, Middle, and End stages), using a PlantVillage dataset labelled with expert guidance. Among these, VGG16 achieved the highest accuracy of 93.5%. Parez et al. (2023) developed E-GreenNet, a customized version of MobileNetV3Small that enhanced its bottleneck layers. This model was trained and tested on PlantVillage, the Data Repository of Leaf Images (DRLI), and a combined PlantComposite dataset to classify leaves as healthy or infected. E-GreenNet outperformed other pre-trained models like VGG16, VGG19, MobileNetV1, and EfficientNetB0, achieving accuracies of 100% on PlantVillage, 96% on DRLI, and 99% on PlantComposite. Doğan (2023) trained MobileNetV3 on the PlantVillage dataset, which was augmented offline to 87,000 RGB images, attaining an accuracy of 99.85%, outperforming ResNet50, EfficientNetB3, and DenseNet121. Pramudhita et al. (2023) compared MobileNetV3Large and EfficientNetB0 on a strawberry leaf dataset with four classes (one healthy and three diseased). By tuning learning rate, epochs, and optimizers, they found that MobileNetV3Large, trained with a learning rate of 0.0001 for 70 epochs using the RMSProp optimizer, achieved the best accuracy of 92.14%.

2.4 Hybrid approaches and ensemble learning

Sahu et al. (2023) developed a hybrid deep learning and machine learning approach using Deep-Dream to classify tomato crops from the PlantVillage dataset. The Deep-Dream network was used to segment lesions in the images for better interpretability. These processed images were then analysed by 24 hybrid models, combining 8 feature extractors (EfficientNet B0-B8) with 3 machine learning classifiers (Random Forest, Stochastic Gradient Boosting, and AdaBoost). After hyperparameter tuning with Optuna, the best model used Deep-Dream with EfficientNet B4 as the feature extractor and AdaBoost as the classifier, achieving 96% accuracy on the dataset and 100% accuracy on a tomato leaf image database from the Indian Agricultural Research Institute. Fenu and Malloci (2023) applied ensemble learning to classify four classes of pear leaf diseases using a private field dataset called DiaMOS Plant (Ganaie et al., 2022). From four networks (EfficientNetB0, InceptionV3, MobileNetV2, and VGG19), the top three were selected to build ensemble models. Three pairs of two-network ensembles were created using bagging and weighted averaging. The best ensemble, combining EfficientNetB0 and InceptionV3, achieved the highest classification accuracy of 91.14%.

2.5 Light-weight architectures

Xu et al. (2022) proposed HLNet (High Speed and Light-weight Network), an optimized version of ShuffleNetV1 was designed to reduce computation time and FLoating-point OPerations (FLOPs). The model incorporated enhanced attention mechanisms, combining channel attention from the SE Network and spatial attention from the convolutional block attention mechanism, to limit computations and boost efficiency. Experiments were conducted on a private dataset combined with PlantVillage and the UC Irvine Machine Learning Repository, covering 28 leaf disease types across 20,490 images from six crop varieties. HLNet achieved 99.86% accuracy with an inference speed of 0.173 s. Thakur et al. (2023) analysed a lightweight Vision Transformer (ViT) (Han et al., 2023; Dosovitskiy et al., 2020) tailored for IoT-based agriculture applications. This model has fewer parameters than many state-of-the-art alternatives and achieved a testing accuracy of 98.86% and precision of 98.90%, outperforming existing models. Table 1 provides a summary of studies related to plant disease identification, highlighting that popular deep learning models like CNN offer various opportunities for enhancing both classification accuracy and execution speed.

Table 1
www.frontiersin.org

Table 1. Summary of plant disease classification studies with methods, datasets, accuracy, technique type, and computational cost.

3 Methodology

The proposed methodology involves designing a customized architecture that integrates the Squeeze-and-Excitation (SE) attention mechanism with a deep neural network. Initially, a backbone model was constructed for classifying various plant disease classes, utilizing convolutional layers, residual/identity blocks, max-pooling, batch normalization, and dense layers. This identity-enhanced backbone is then combined with the attention mechanism to enhance performance and stability in multi-class classification. This approach facilitates accurate differentiation among multiple plant categories, including both healthy and diseased conditions. Figure 1 depicts the backbone framework employed for the multi-class classification of plant images.

Figure 1
Diagram illustrating the architecture of a proposed CNN-SEEI network. Part (a) shows the flow from an input image through various layers such as convolution, identity blocks with squeeze and excitation, and dense layers. Part (b) details the squeeze-and-excitation block, highlighting global average pooling, fully connected layers, ReLU activation, and sigmoid functions. Part (c) depicts the identity block with a sequence of convolution layers and squeeze-and-excitation operation.

Figure 1. (a) The framework of the proposed CNN-SEEIB for plant disease classification. (b) Squeeze – and – Excitation (SE) Block. (c) Identity block with Squeeze – and – Excitation (SE).

3.1 CNN-SEEIB backbone architecture

The CNN-SEEIB backbone is a simple variant of the conventional CNN architecture, modelled to incorporate squeeze-and-excitation attention mechanisms. It consists of several fundamental components of CNNs: the input layer, convolutional blocks, identity blocks, max pooling, batch normalization, and activation layers.

3.1.1 Input layer

The input layer accepts a three-channel image which is resized into patches of 224 × 224 × 3 pixels to standardize the input dimensions.

3.1.2 Convolutional blocks

Each convolutional block performs 3 × 3 convolutions to extract local features from the input feature maps, followed by max pooling with a 2 × 2 window to reduce spatial dimensions and computational speed. This process enhances feature representation by transforming local features into more abstract global features as the network deepens. In CNN-SEEIB, there are four convolutional blocks that form the primary feature extractor.

3.1.3 Identity blocks

The identity block is employed as a bottleneck design with a 1 × 1 convolution to reduce channel dimensions, then a 3 × 3 convolution processes these features, and finally another 1 × 1 convolution to restore the original channel size. The input of the block is then added element-wise to the output of the last convolution, facilitating the retention of essential features and avoid vanishing gradient issues through skip connections. This mechanism helps to maintain gradient flow and supports effective learning of complex features. CNN-SEEIB includes four such identity blocks interspersed with convolutional blocks.

3.1.4 Max pooling layer

Max pooling, serves to down-sample input feature maps, enhancing computational efficiency and mitigating overfitting risks. It is typically applied post-convolutional layers to reduce spatial dimensions. During max pooling, each channel of the input feature map is processed independently using a pooling window, typically of size 2 × 2, where the maximum value within the region is extracted, discarding others. This process is iterated with a defined stride, dictating the window’s movement across the input feature map. Mathematically, the max pooling operation for a window of size x × x where stride of s can be expressed as given in Equation 1:

M ijk = max p = 0 x 1 max q = 0 x 1 ( N ( si + p ) , ( sj + q ) , k )     (1)

where M ijk is the max pooling layer output at position (i,j) in channel k, and N ( si + p ) , ( sj + q ) , k refers the input value at position ( si + p , sj + q ) in channel k as stride s and x are the pooling window size.

3.1.5 Batch normalization

Batch normalization normalizes the input activations across each mini-batch by subtracting the batch mean and dividing by the batch standard deviation, then applying learnable scaling and offset parameters. This normalization stabilizes and speeds up training while acting as a regularizer to reduce overfitting.

3.1.6 Activation functions

The ReLU activation function, defined as max (0, x), introduces non-linearity after each convolution and fully connected layer, enabling the network to learn complex patterns. The final classification layer employs the SoftMax activation to output class probabilities.

3.1.7 Global average pooling

Global Average Pooling (GAP) condenses the spatial dimensions of feature maps and delivers a concise feature representation. It includes computing the average of each feature map, yielding a single value for each channel, which can effectively reduce the network’s parameters and computational load. When applied to a feature map F of spatial position (i, j) with dimensions H × W × C (where H is height, W is width and C is the number of channels), the GAP operation generates a 1D output vector G with length C. The computation for the c th channel in the output vector G is given as Equation 2:

G c = 1 HXW i = 1 H j = 1 W F i , j , c     (2)

This model used a global average pooling after the identity layer. GAP helps in making the network translation-invariant by summarizing the features in each channel, irrespective of their spatial location resulting in compact and informative feature representations.

Table 2 details the summary of the layers and parameters in the backbone CNN architecture. In CNN-SEEIB, the interaction between convolutional and identity blocks, enhanced by batch normalization and ReLU activation, ensures effective feature extraction and gradient flow. The model balances depth and computational efficiency with a total of approximately 3.34 million parameters, of which around 3.33 million are trainable, structured across 4 convolutional and 4 identity blocks.

Table 2
www.frontiersin.org

Table 2. Layer-wise summary of the backbone CNN architecture with output shapes, specifications, and parameter count.

3.2 Improvisations through squeeze and excitation block

The utility of attention mechanisms enhances image recognition models by focusing on salient features within images, machine translation systems by highlighting relevant parts of a sentence. By enabling models to adapt their focus based on the input data, attention mechanisms improve the interpretability of neural network decisions and enhance the model’s reasoning process, boosting performance across diverse tasks (Hu et al., 2018; Woo et al., 2018). The SE is an attention mechanism used to adaptively reconfigure feature maps to enhance the representational power of CNNs. The SE block consists of two main functions: squeeze operation and excitation operation.

3.2.1 Squeeze operation

To effectively capture inter-channel dependencies, the output features of each channel are analysed, as individual units within the transformation output L lack the capacity to extract global contextual features due to the constraints imposed by their limited receptive fields. Given a set of input feature maps I f R HXWXC , where each I f c R HXW , it becomes essential to squeeze spatial information across the entire feature map into a channel-wise representation. To address this, a global average pooling operation which is simple and effective is utilized to generate statistics for each channel. This operation compresses the spatial dimensions height H and width W of the input feature maps, resulting in a channel descriptor s R c Each component s C of this descriptor reflects the aggregated spatial response of the corresponding channel c and is derived as shown in Equation 3.

s C = F Squeeze ( I f c ) = 1 H X W p = 1 H q = 1 W I f c ( p , q )     (3)

3.2.2 Excitation operation

To enhance the model’s ability to adaptively recalibrate feature representation, the excitation phase is incorporated after the squeeze operation. The squeeze operation captures global spatial information by condensing each channel’s activation into a single scalar. The excitation phase utilizes this global context to effectively model channel inter-dependencies capturing the non-linear and non-mutually exclusive relationships between channels. To achieve this, a gating mechanism equipped with a sigmoid activation function is utilized. The excitation function is mathematically expressed in Equation 4:

O = F Excitation ( s , T ) = σ ( gating ( s , T ) ) = σ ( T 2 ReLU ( T 1 s ) )     (4)

Here, s R c represents the channel-wise descriptors obtained from the squeeze operation, T 1 R c r × c and T 2 R c × c r are the learnable weights of two fully connected (FC) layers, and r is the reduction ratio which controls the dimentionality compression and bottleneck capacity. The ReLU function introduces non-linearity, while the sigmoid function σ ensures, the output values (attention weights) lie within the range [0, 1].

This excitation mechanism consists of a lightweight two-layer MLP (multi-layer perceptron). The first FC layer reduces the channel dimensionality to c r , acting as a compression step, and is followed by a ReLU activation. The second FC layer restores the original channel dimension c , thereby resulting in the attention weights for recalibrating the original feature maps.

The recalibration is performed via channel-wise multiplication, where each channel of the transformation output is scaled by the corresponding excitation weight. This operation is described in Equation 5:

O Ic ˜ = F scale ( O Ic , z c ) = z O Ic     (5)

In this equation, O Ic R H X W denotes the feature map for channel c from the transformation output and z c [ 0 , 1 ] is the scalar excitation weight corresponding to that channel. The function F scale represents an element-wise channel scaling operation, which selectively emphasizes or suppresses channel responses based on their contextual importance.

As illustrated in Figure 1b, the complete Squeeze-and-Excitation (SE) Attention Block provides an efficient mechanism for dynamic channel-wise feature recalibration, enhancing representational capacity with minimal additional computational overhead. The proposed hybrid architecture integrates convolutional blocks with identity blocks (Figure 1c), enabling deeper models while mitigating the vanishing gradient problem through skip connections. Concatenation with max-pooling operations facilitates multi-scale feature representation across different levels of abstraction, thereby enhancing the model’s capability to extract rich and discriminative features. Batch normalization is utilized to stabilize and accelerate training by normalizing intermediate features thus improving convergence and generalization. To reduce overfitting and computational complexity, global average pooling is incorporated prior to the dense layers for effectively decreasing the number of parameters.

The integration of the Squeeze-and-Excitation mechanism enhances the model’s representational power by modelling channel-wise interdependencies. The squeeze operation utilizes global average pooling to capture spatially global information, while the excitation operation computes adaptive channel weights that are used to recalibrate feature maps. This selective importance on informative channels allows the network to focus on the most task-relevant features, effectively suppressing redundant or noisy information. By embedding the SE block, the model achieves superior feature discrimination, leading to improved classification accuracy and robustness. This attention-driven enhancement makes the model more resilient to variations in input data, which is evident from the improved experimental metrics compared to baseline architectures.

3.3 Computational complexity

In general, the overall runtime of the convolution blocks in CNN is linear, and therefore for classifying images with n pixels, the computational complexity of the feature extractor blocks is O ( n ) . This is due to the fact that the complexity of the convolution operation is O ( n ) and the other operations like pooling and SE do not take more than the linear time. As given in Lux (n.d.), if the extracted feature map is of dimension N , the dense neural network layer (i.e., feed forward neural network) has a computational complexity of O ( N 4 ) . Therefore, the complexity of the proposed CNN-SEEIB is O ( N 4 + n ) O ( N 4 ) since n N 4 . Note that the computational complexity of the proposed architecture is comparable to that of the feed forward neural network.

4 Results and discussion

This section details (i) Dataset Description, which outlines the dataset’s characteristics, (ii) Experimentation, detailing the experimental setup, training parameters, and evaluation criteria, and (iii) Result Analysis, which interprets the obtained outcomes, including training and validation results, testing outcomes, inferencing outcomes, and comparisons with other models.

4.1 Dataset description and processing

This proposed work analysed the widely used PlantVillage dataset, renowned for its extensive and diverse collection of plant leaf images (Mohanty et al., 2016). The dataset contains a total of 54,305 images representing 38 distinct classes, which include 14 crop species such as apple, blueberry, cherry, grape, orange, peach, pepper, potato, raspberry, soy, squash, strawberry, and tomato. These 38 classes are categorized into 17 fungal diseases, 4 bacterial diseases, 2 diseases caused by mold (oomycetes), 2 viral diseases, 1 disease caused by a mite, and 12 classes representing healthy crops. Figures 2, 3 illustrate the distribution of classes across the dataset and shows the representative leaf samples from each of the 38 categories. Prior to training, the dataset undergoes a preprocessing process involving the removal of duplicate entries and validation of file formats. All images are then resized to 224 × 224 × 3 and the dataset is split into training (80%), validation (10%), and testing (10%) subsets to support model training and evaluation.

Figure 2
Bar chart titled

Figure 2. Class wise distribution of the PlantVillage dataset.

Figure 3
A grid of various leaves showing different shapes, sizes, and colors, some with visible damage or spots. The leaves are arranged in several rows on a neutral gray background.

Figure 3. Sample leaf images from each of the 38 classes from the PlantVillage dataset.

4.2 Experimental setup

The training, testing, and inference experiments were conducted using Google Collaboratory with a T4 cloud GPU backend. The implementation was carried out in Python 3.10, utilizing TensorFlow-Keras 2.15.0 as the deep learning framework. Model inference was performed on a local machine equipped with an Intel (R) Core (TM) i7 CPU, 16GB RAM, and an NVIDIA GeForce GTX 1650 Ti GPU with 4GB of dedicated memory. To monitor hardware performance metrics such as CPU and GPU utilization and power consumption, HWiNFO software was employed. The software tracks four key parameters: total CPU usage, GPU D3D usage, CPU package power, and GPU power. For evaluation, 100 samples of each parameter were collected over a 3-min interval, and the average values were calculated. Additionally, the net_flops (model) function from the net_flops Keras library was used to compute the FLOPS (floating point operations per second) of the model.

4.3 Hyperparameter tuning

The CNN-SEEIB model was fine-tuned to improve its performance by adjusting several key hyperparameters, like dropout rate, batch size, epochs, learning rate, and gradient optimizer. This study employed the Adam optimizer (adaptive moment estimation) (Kingma and Ba, 2015), which is known for its efficiency and low memory requirements. Adam combines the benefits of gradient descent with momentum and the RMSProp optimizer (GeeksforGeeks, n.d.), enhances the plant leaf classification. Table 3 details the list of hyperparameters that are tuned to attain better results.

Table 3
www.frontiersin.org

Table 3. Hyperparameters for the proposed CNN-SEEIB model.

The loss during Back Propagation is computed using Categorical Cross entropy loss function as defined as Equation 6.

L ( y , y ̂ ) = 1 N i = 1 N k = 1 K y i , k log ( y ̂ i , k )     (6)

where y is the ground truth distribution (one-hot encoded labels), y ̂ is the predicted probability distribution (the output of the SoftMax activation function), y i , k is the true label for the ith sample in class k, and y ̂ i , k is the predicted probability for the ith sample belonging to class k.

4.4 Evaluation metrics

Evaluation metrics play a pivotal role in assessing the efficacy of the developed and trained model. In this study, the evaluation metrics are categorized into two groups: (i) testing metrics and (ii) inferencing metrics.

4.4.1 Metrices for testing

The classification performance of the proposed model was evaluated based on four key metrics derived from the confusion matrix: accuracy, precision, recall and F1-score. While accuracy indicates the overall percentage of correct predictions, there might be instances where false predictions have different implications. In such scenarios, precision, recall and F1-score provide a more comprehensive evaluation of the model.

True Positives (TP) refer to the number of instances where the model correctly identifies diseased leaves as belonging to a disease class, i.e., actual positives correctly predicted as positives. In the context of the PlantVillage dataset, this means the model accurately detects the presence of a specific leaf disease. True Negatives (TN) denote the number of healthy leaf samples that are correctly identified as healthy, i.e., actual negatives predicted as negatives. False Positives (FP) represent healthy leaves that are incorrectly classified as diseased, i.e., actual negatives mistakenly predicted as positives. False Negatives (FN) indicate diseased leaves that the model incorrectly classifies as healthy, i.e., actual positives predicted as negatives.

The definitions of the evaluation metrics are given in Equations 710:

Accuracy = TP + TN TP + TN + FP + FN × 100     (7)
Precision ( P ) = TP TP + FP     (8)
Recall ( R ) = TP TP + FN     (9)
F 1 Score = 2 × P × R P + R     (10)

4.4.2 Metrics for inferencing

To assess the computational capabilities, the following inference metrics were considered: (i) number of parameters and model size, (ii) inference time and (iii) frames per second.

4.4.2.1 Number of parameters and model size

These metrics depend upon the architecture of the model based on the number of trainable and non-trainable parameters, and the memory occupied by these parameters during when the model is loaded.

4.4.3 Inference time

It refers to the duration of time required for a model to process single input and generate predictions. It measures the computational efficiency of the model.

4.4.4 Frames per second

Frames Per Second (FPS) measures the number of frames a model can process and render per second. A higher FPS indicates smoother performance and better visual quality and is a critical factor in real-time applications.

4.5 Performance evaluation

During the training phase, the proposed CNN-SEEIB architecture, along with eight different pre-trained models, was trained on the dataset for 50 epochs using the same set of hyperparameters. The pre-trained models used include VGG16, ResNet50, EfficientNetB0, DenseNet121, MobileNetV2, InceptionV3, Inception-ResNetV2, and XceptionNet. Table 4 presents the accuracy of each model along with the input dimensions for both training and validation.

Table 4
www.frontiersin.org

Table 4. Training and validation accuracy comparison of the proposed CNN-SEEIB model and pre-trained models.

All models were trained to achieve optimal accuracy on the training dataset, incorporating regularization techniques to prevent overfitting. The selection of models represents a range of architectural improvements ranging from depth-scaling design of VGG16 to the more intricate scaling strategies employed by EfficientNet-B0. ResNet-50 and DenseNet-121 were chosen for their robust deep architectures leveraging residual and dense connections, respectively. MobileNet-V2 provides a low-weight solution that is appropriate for data with limited constraints, making it perfect for applications that require real-time processing. Inception-V3 and Inception-ResNetV2 utilise multi-scale processing and hybrid architectures to successfully capture a wide range of features. These models encompass early stopping, to halt training upon validation accuracy saturation, and model checkpoints to save the model weights with the highest accuracy in the validation set. As shown in Table 4, the proposed CNN-SEEIB model achieved the highest validation accuracy among all compared models, whereas VGG16 recorded the lowest. In addition to the enhanced proposed model, only InceptionV3 reached a peak accuracy of 99.89%. Figures 4, 5 illustrate the training loss and accuracy curves for each pre-trained model as well as the CNN-SEEIB model.

Figure 4
Line graph comparing the loss over 30 epochs for various neural networks, including VGG16, ResNet50, EfficientNet B0, among others. Most models show a rapid decrease in loss initially, stabilizing after about 10 epochs. The legend on the right lists the models with corresponding colors.

Figure 4. Training loss curves for pre-trained models and the proposed CNN-SEEIB model.

Figure 5
Line graph showing model accuracy over 30 epochs for various neural networks. All models quickly reach high accuracy, converging near 1.0. Legends detail models including VGG16, ResNet50, EfficientNet B0, MobileNetV2, and others, each represented by different colored lines.

Figure 5. Training accuracy curves for pre-trained models and the proposed CNN-SEEIB model.

Table 5 presents the comparative results of multi-class classification performance on the test dataset between the proposed CNN-SEEIB model and the pre-trained models. The CNN-SEEIB model outperformed all others, achieving an accuracy of 99.79%, with precision, recall, and F1-score values of 0.9972, 0.9970, and 0.9971, respectively. The incorporation of the SE attention mechanism enhanced the model’s ability to focus on critical features. Among the pre-trained models, DenseNet121 demonstrated the best performance across all evaluation metrics, while VGG16 exhibited the lowest.

Table 5
www.frontiersin.org

Table 5. Performance comparison of models based on accuracy, precision, recall, and F1-score.

Models that deliver high classification performance while maintaining low resource consumption are considered optimal. Through inferencing, the resource utilization metrics such as model size, inference time, and frames per second are measured and are evaluated. Table 6 show the results of this comparison based on inference performance. The measurement of the metrics has been carried out in a local runtime to avoid latency errors in measurement. The model size and parameters are obtained through model statistics when the model is loaded while the inference time and frames per second metrics are measured as average over time when the model is evaluated in real-time. It is clear that mobilenetV2 has the lowest inference time, Inception-ResnetV2 has the highest number of parameters and memory, and DenseNet121 takes the largest inference time. The proposed model improvised through SE mechanism stands next to MobileNetV2 in inference time and differs only by a small margin. The proposed CNN-SEEIB has the smallest size and number of parameters and performs faster than most of the pre-trained models except for MobileNetV2.

Table 6
www.frontiersin.org

Table 6. Inference performance and resource utilization comparison of models.

Resource utilization and power consumption metrics are vital for assessing the feasibility of deploying the CNN-SEEIB model on edge devices (Chen and Ran, 2019). To evaluate system resource usage, CPU and GPU utilization percentages serve as key metrics, while the power consumption of both CPU and GPU (measured in watts) indicates the load the model imposes on the hardware. Although these metrics do not directly impact the classification task, they influence the hardware or system where the model operates. Figures 6, 7 illustrate the resource utilization details for the proposed CNN-SEEIB and its backbone CNN architecture. The results show that integrating the SE mechanism in CNN-SEEIB leads to a 0.35% increase in CPU utilization and a 0.4% increase in GPU utilization. Similarly, CPU power consumption rises by approximately 7%, and GPU power consumption by about 6%. However, these increases are justified by the notable gains in classification accuracy achieved by the model. Moreover, the efficient design of CNN-SEEIB keeps overall GPU utilization relatively low by optimizing resource use and minimizing computations.

Figure 6
Bar chart comparing CPU and GPU utilization for Backbone CNN without SE and Proposed CNN-SEEIB. CPU utilization is slightly higher in the proposed model. GPU utilization is roughly equal in both models. Utilization is measured as a percentage up to 12%.

Figure 6. CPU and GPU utilization between the backbone CNN (without SE) and the proposed CNN-SEEIB.

Figure 7
Bar chart comparing power consumption of CPU and GPU using Backbone CNN (without SE) and Proposed CNN-SEEIB. CPU power consumption is higher for both, with the Proposed CNN-SEEIB slightly exceeding the Backbone CNN. GPU power consumption is lower for both, with the Proposed CNN-SEEIB slightly less than the Backbone CNN.

Figure 7. CPU and GPU power consumption between the backbone CNN (without SE) and the proposed CNN-SEEIB.

In resource-constrained environments, these modest trade-offs are reasonable, given the enhanced decision-making accuracy of the model. Potential further optimizations, such as model quantization or pruning, could reduce resource consumption without sacrificing performance, improving its suitability for edge deployment. Additionally, as energy-efficient hardware becomes more accessible, the practical impact of such increases will diminish. These considerations underscore the balance between accuracy and efficiency, making CNN-SEEIB a scalable and sustainable solution for various real-world applications.

4.6 Visualization of activation maps

Activation Maps offer valuable insights into the features captured by the model internally, shedding light on its function and behavior. As features flow from the input to the output layer, they become increasingly global and abstract, a process comprehended by the dense layers for accurate classification. Figure 8 visually represents the intermediate feature maps of the proposed CNN-SEEIB model, illustrating the progression of features through the network.

Figure 8
Grid of leaf images with a color scale from dark blue to yellow. Each image displays varying patterns and hues, analyzed on a vertical color bar from 0.02 to 0.05 on the right.

Figure 8. Intermediate feature maps obtained from the CNN-SEEIB model.

4.7 Ablation study

To analyze the impact of the SE mechanism on the proposed CNN-SEEIB model, an ablation study was conducted by assessing the model’s performance without the SE module. Figure 9 presents the accuracy curves for CNN-SEEIB both with and without SE during training and validation. It is observed that, without SE, the validation accuracy lags behind the training accuracy even at the 50th epoch. In contrast, with the SE mechanism included, the validation accuracy aligns closely with the training accuracy from the 35th epoch onward. This demonstrates that the SE module significantly contributes to the accuracy improvement of the proposed CNN-SEEIB model.

Figure 9
Line graph showing accuracy versus epoch for CNN-SEEIB models. Four lines represent: CNN-SEEIB with SE Attention (Training) in blue, Validation in green; without SE Attention (Training) in red, Validation in purple. Both models reach high accuracy around epoch 30.

Figure 9. Model accuracy of CNN-SEEIB with and without squeeze and excitation (SE) attention.

4.8 Performance comparison with the state-of-the-art

The proposed model in the study is further compared with several other architectures from existing studies to highlight its significance and performance as shown in Table 7.

Table 7
www.frontiersin.org

Table 7. Comparison of accuracy results with the existing state-of-the-art models.

Each model achieved peak accuracies ranging from 99.41 to 99.79%, demonstrating their effectiveness in accurately identifying and classifying plant diseases. The proposed CNN-SEEIB, outperformed all others, reaching the highest accuracy of 99.79%. The model proposed in these studies (Hassan et al., 2021; Brahimi et al., 2018) diverges from traditional transfer learning approaches as it leverages the custom hybrid architecture incorporated with the SE mechanism that significantly enhances the model’s capability to discern intricate features and patterns within plant images. Despite the sophisticated design, the model remains lightweight, offering a remarkable balance between performance and computational efficiency. The high accuracy attained highlights the model’s potential for practical deployment in agricultural settings, where accurate and efficient disease identification is crucial for ensuring crop health and yield.

4.9 Validation of robustness and generalizability

Despite the homogeneous background and controlled conditions of the PlantVillage dataset, it has been widely utilized in numerous studies as a benchmark for plant disease classification. The class imbalance issue with this dataset is a potential challenge as certain disease classes contain more samples than others, potentially resulting in biased predictions. The datasets may not fully reflect actual agricultural contexts because they were collected in lab settings with uniform backgrounds and ideal illumination. Furthermore, the dataset has noise associated with the labels, which allows DL models to alter prediction biases rather than learning the disease features itself. This could potentially impact the performance of the models in real-world scenarios. To validate the robustness and the generalization of the proposed CNN-SEEIB model, the experiment is repeated on the potato leaf disease dataset (Rashid et al., 2021). Early detection of potato leaf diseases is challenging due to the diversity in crop species, disease symptoms, and varying climatic conditions, which add complexity to the detection process (Rashid et al., 2021). The potato leaf disease dataset comprises of 4,062 images collected from the Central Punjab region of Pakistan.

This dataset comprises a diverse collection of images categorized into three distinct classes: early blight, late blight, and healthy each representing a specific condition affecting potato crops. Early blight, caused by the fungus Alternaria solani, typically affects older plants and is also known as target spot. Late blight, caused by Phytophthora infestans, is a severe disease that can lead to rapid crop failure without proper control measures. In contrast, healthy potatoes are nutrient-rich, providing essential elements like vitamin C and potassium. The dataset was divided into 3,251 training images, 416 testing images, and 405 validation images. The sample potato leaf images used for training and validation is shown in Figure 10.

Figure 10
Three images of potato leaves side by side. The first leaf shows early blight with dark spots. The second leaf displays late blight with darker patches. The third leaf is healthy and green.

Figure 10. Sample potato leaf disease images.

The proposed CNN-SEEIB was used on this dataset to classify three categories: Early Blight, Healthy, and Late Blight. It achieved an accuracy of 97.77%, precision of 0.9746, recall of 0.9808 and an F1 score of 0.9773. Figures 11, 12 shows the accuracy and loss plot and Figure 13 shows the confusion matrix obtained for the potato leaf disease detection. The confusion matrix shows that Healthy leaves (Class 0) are most often misclassified, with 5 instances as Early Blight (Class 1) and 2 as Late Blight (Class 2), likely due to subtle early-stage symptoms resembling healthy leaves, while Early Blight is correctly classified, indicating its discrete visual markers like well-defined spots that are easily detected. Late Blight (Class 2) has minor errors, with 1 misclassified as Healthy and 1 as Early Blight, possibly due to overlapping features like lesion patterns or discoloration. These misclassifications show the CNN-SEEIB model has to be improved with diseases sharing visual traits, such as spot shapes or color changes. Enhancing the model’s ability to identify fine-grained differences, like texture or spot distribution, an improve the performance on the potato leaf disease dataset. The performance of the proposed CNN-SEEIB model on a dataset distinct from PlantVillage shows its potential for generalization across diverse datasets and real-world scenarios. Figure 14 details ROC Plot for CNN-SEEIB’s performance on potato leaf disease dataset. The inner layers of the proposed CNN-SEEIB model generate visual feature representations, as shown in Figure 15, using Grad-CAM. These Grad-CAM-based visualizations provide a detailed understanding of the discriminative regions the model focuses on while analyzing the potato leaf disease dataset. During training, each layer produces feature maps that capture specific patterns related to disease symptoms. This visualization technique is used in highlighting which regions of the potato leaf images influence the model’s predictions, showing critical insights into the model’s decision-making process and interpretability.

Figure 11
Line graph titled

Figure 11. Accuracy plot for CNN-SEEIB’s performance on potato leaf disease dataset.

Figure 12
Line graph titled

Figure 12. Loss plot for CNN-SEEIB’s performance on potato leaf disease dataset.

Figure 13
Confusion matrix with three classes. True labels are on the Y-axis and predicted labels on the X-axis. Values: 155 (top-left), 102 (center), and 139 (bottom-right). Color scale indicates intensity.

Figure 13. Confusion matrix for CNN-SEEIB’s performance on potato leaf disease dataset.

Figure 14
ROC curve illustrating the true positive rate versus false positive rate for different potato conditions: early blight (blue, AUC = 1.00), late blight (orange, AUC = 0.99), and healthy (green, AUC = 0.98). A diagonal dashed line represents random chance.

Figure 14. ROC plot for CNN-SEEIB’s performance on potato leaf disease dataset.

Figure 15
First row shows a healthy leaf, a blank Grad-CAM heatmap, and a superimposed image of the leaf. Second row shows an early blight-affected leaf, its Grad-CAM heatmap with varied colors, and the superimposed image highlighting affected areas. Third row displays a late blight-affected leaf, its colorful Grad-CAM heatmap, and the superimposed image emphasizing the blight regions.

Figure 15. Intermediate feature representation of potato leaf disease.

5 Conclusion

This paper presents a custom CNN-based model, the CNN-SEEIB that integrates the benefits of accurate and lightweight architectures to automatically classify various plant leaf diseases from images. Initially, a custom backbone CNN architecture was designed, featuring a reduced number of parameters. With this CNN acting as the backbone, attention-based optimizations are incorporated using squeeze and excitation resulting in a unique model with superior classification metrics without significantly increasing parameter count. The experimentation on the PlantVillage dataset shows that the proposed CNN-SEEIB model attained a classification accuracy of 99.79%, precision of 0.9970, recall of 0.9972, and an F1 score of 0.9971. Real-time inferencing revealed an inference time per image of 64 ms. Compared to several transfer learning models, the proposed model outperforms all others in accuracy, precision, recall, and F1-score. With the squeeze-and-excitation model, CNN-SEEIB’s inference time is almost equal to that of MobileNetV2. Also, CNN-SEEIB has the smallest model size and lowest parameter count, while the execution is faster than most pre-trained models except MobileNetV2. To further assess its robustness and generalizability, the model was also evaluated on the potato leaf disease dataset, achieving an accuracy of 97.77%, precision of 0.9746, recall of 0.9808, and F1-score of 0.9773. These results confirm that CNN-SEEIB performs effectively across different datasets and real-world conditions.

Recommendations for further research are as follows:

• Utilizing additional regularization techniques such as bagging and boosting ensemble methods to enhance the generalization capability of the model.

• Extending the classification model to incorporate disease detection and identification functionalities by implementing object detection and segmentation techniques to identify infected regions.

• Augmenting the existing dataset using GANs, NST, and BIM techniques, or combining multiple datasets to increase the number of samples, for reducing the bias and improving model robustness.

• Expanding the number of classes to include a wider range of diseases affecting both leaves, stems and roots, and employing multi-label classification approaches to identify multiple diseases simultaneously.

• Combining other ML techniques for classification, and semi-supervised learning methods to leverage unlabelled data, either imported or self-collected, for model enhancement.

• Exploring various attention mechanisms like self-attention, multi-head attention, soft and hard attention and cross-attention can help recognizing the approach to improve leaf disease detection performance by concentrating on related features and precise patterns.

• Training the proposed model across various other datasets (say, medical images) can help to evaluate its effectiveness and robustness.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/datasets/mohitsingh1804/plantvillage.

Author contributions

SK: Data curation, Formal analysis, Methodology, Resources, Software, Validation, Writing – original draft, Writing – review & editing. RC: Formal analysis, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. SN: Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. LJ: Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ahmad, M., Abdullah, M., Moon, H., and Han, D. (2021). Plant disease detection in imbalanced datasets using efficient convolutional neural networks with stepwise transfer learning. IEEE Access 9, 140565–140580. doi: 10.1109/access.2021.3119655

Crossref Full Text | Google Scholar

Alzubaidi, L., Zhang, J., Humaidi, A. J., al-Dujaili, A., Duan, Y., al-Shamma, O., et al. (2021). Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J. Big Data 8:53. doi: 10.1186/s40537-021-00444-8

PubMed Abstract | Crossref Full Text | Google Scholar

Ashurov, A. Y., Al-Gaashani, M. S. A. M., Samee, N. A., Alkanhel, R., Atteia, G., Abdallah, H. A., et al. (2025). Enhancing plant disease detection through deep learning: a depthwise CNN with squeeze and excitation integration and residual skip connections. Front. Plant Sci. 15:1505857. doi: 10.3389/fpls.2024.1505857

PubMed Abstract | Crossref Full Text | Google Scholar

Assaduzzaman, M., Bishshash, P., Nirob, M. A. S., Marouf, A. A., Rokne, J. G., and Alhajj, R. (2025). XSE-TomatoNet: an explainable AI based tomato leaf disease classification method using EfficientNetB0 with squeeze-and-excitation blocks and multi-scale feature fusion. MethodsX 14:103159. doi: 10.1016/j.mex.2025.103159

PubMed Abstract | Crossref Full Text | Google Scholar

Balafas, V., Karantoumanis, E., Louta, M., and Ploskas, N. (2023). Machine learning and deep learning for plant disease classification and detection. IEEE Access 11, 114352–114377. doi: 10.1109/ACCESS.2023.3324722

Crossref Full Text | Google Scholar

Batchuluun, G., Nam, S. H., and Park, K. R. (2022). Deep learning-based plant-image classification using a small training dataset. Mathematics 10:3091. doi: 10.3390/math10173091

Crossref Full Text | Google Scholar

Bhuyan, P., Singh, P. K., Das, S. K., and Kalla, A. (2023). SE_SPnet: rice leaf disease prediction using stacked parallel convolutional neural network with squeeze-and-excitation. Expert. Syst. 40:e13304. doi: 10.1111/exsy.13304

Crossref Full Text | Google Scholar

Brahimi, M., Arsenovic, M., Laraba, S., Sladojevic, S., Boukhalfa, K., and Moussaoui, A. (2018). “Deep learning for plant diseases: detection and saliency map visualisation” in Human and machine learning Ed. Jianlong Zhou and Fang Chen. (Cham, Switzerland: Springer International Publishing), 93–117.

Google Scholar

Chen, J., and Ran, X. (2019). Deep learning with edge computing: a review. Proc. IEEE 107, 1655–1674. doi: 10.1109/JPROC.2019.2921977

Crossref Full Text | Google Scholar

Chen, J., Zhang, D., Suzauddola, M., Nanehkaran, Y. A., and Sun, Y. (2021). Identification of plant disease images via a squeeze-and-excitation MobileNet model and twice transfer learning. IET Image Process. 15, 1115–1127. doi: 10.1049/ipr2.12090

Crossref Full Text | Google Scholar

Doğan, Y. (2023). Comparative analysis of CNN models for plant disease classification using augmented PlantVillage dataset. AS-Proceedings 1, 1075–1079. doi: 10.59287/as-proceedings.845

Crossref Full Text | Google Scholar

Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, W., Zhai, X., Unterthiner, T., et al. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv. doi: 10.48550/arXiv.2010.1192

Crossref Full Text | Google Scholar

Fenu, G., and Malloci, F. M. (2023). Classification of pear leaf diseases based on ensemble convolutional neural networks. Agri 5, 141–152. doi: 10.3390/agriengineering5010009

Crossref Full Text | Google Scholar

Ganaie, M. A., Hu, M., Malik, A. K., Tanveer, M., and Suganthan, P. N. (2022). Ensemble deep learning: A review. Eng. Appl. Artif. Intell. 115:105151. doi: 10.1016/j.engappai.2022.105151

Crossref Full Text | Google Scholar

GeeksforGeeks. Available online at: https://www.geeksforgeeks.org/adam-optimizer/ (last accessed date: 30-August-2024)

Google Scholar

Guo, J. (2023). Classification of crop disease images based on convolutional neural network. Adv. Comput. Commun. 4, 205–209. doi: 10.26855/acc.2023.06.020

Crossref Full Text | Google Scholar

Han, K., Wang, Y., Chen, H., Chen, X., Guo, J., Liu, Z., et al. (2023). A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 45, 87–110. doi: 10.1109/tpami.2022.3152247

PubMed Abstract | Crossref Full Text | Google Scholar

Hassan, S. M., Maji, A. K., Jasiński, M., Leonowicz, Z., and Jasińska, E. (2021). Identification of plant-leaf diseases using CNN and transfer-learning approach. Electronics 10:1388. doi: 10.3390/electronics10121388.s

Crossref Full Text | Google Scholar

Hu, J., Shen, L., and Sun, G. (2018). Squeeze-and-excitation networks. Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 7132–7141. doi: 10.1109/CVPR.2018.00745

Crossref Full Text | Google Scholar

Imanulloh, S. B., Muslikh, A. R., and Setiadi, D. R. I. M. (2023). Plant diseases classification based leaves image using convolutional neural network. J. Comput. Theor. Appl. 1, 1–10. doi: 10.33633/jcta.v1i1.8877

Crossref Full Text | Google Scholar

Kingma, D. P., and Ba, L. J.. Adam: a method for stochastic optimization. In Proc. International Conference on Learning Representations (ICLR), (2015).

Google Scholar

Maeda-Gutiérrez, V., Galván-Tejada, C. E., Zanella-Calzada, L. A., Celaya-Padilla, J. M., Galván-Tejada, J. I., Gamboa-Rosales, H., et al. (2020). Comparison of convolutional neural network architectures for classification of tomato plant diseases. Appl. Sci. 10:1245. doi: 10.3390/app10041245

Crossref Full Text | Google Scholar

Majji, V. A., and Kumaravelan, G. (2021). Detection and classification of plant leaf disease using convolutional neural network on Plant Village dataset. Int. J. Innov. Res. Appl. Sci. Eng. (IJIRASE) 4, 931–935. doi: 10.29027/IJIRASE.v4.i11.2021.931-935

Crossref Full Text | Google Scholar

Mohanty, S. P., Hughes, D. P., and Salathé, M. (2016). Using deep learning for image-based plant disease detection. Front. Plant Sci. 7:215232. doi: 10.3389/fpls.2016.01419

Crossref Full Text | Google Scholar

Naik, B., Nageswararao, R., Malmathanraj, R., and Palanisamy, P. (2022). Detection and classification of chilli leaf disease using a squeeze-and-excitation-based CNN model. Ecol. Inform. 69:101663. doi: 10.1016/j.ecoinf.2022.101663

Crossref Full Text | Google Scholar

Naresh, V., Yogeswararao, G., Nageswararao Naik, B., Malmathanraj, R., and Palanisamy, P. (2024). Empirical analysis of squeeze and excitation-based densely connected CNN for chili leaf disease identification. IEEE Trans. Artif. Intell. 5, 1681–1692. doi: 10.1109/TAI.2024.3364126

Crossref Full Text | Google Scholar

National Institute of Food and Agriculture. (2022). Researchers are helping protect crops from pests. U.S. Department of Agriculture. Available online at: https://www.nifa.usda.gov/about-nifa/blogs/researchers-helping-protect-crops-pests

Google Scholar

Nikhileswar, G. S., Pampana, S. S., Anbarasi, L. J., Bhavana, C. L., Nanditha, R. L., and Krishna, K. C. (2024). PlantLDNet: disease detection in leaf images using deep learning models. 2024 5th International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV). IEEE.

Google Scholar

Pan, S. J., and Yang, Q. (2010). A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22, 1345–1359. doi: 10.1109/TKDE.2009.191

Crossref Full Text | Google Scholar

Panchal, A. V., Patel, S. C., Bagyalakshmi, K., Kumar, P., Khan, I. P., and Soni, T. (2021). Image-based plant diseases detection using deep learning. Mater Today Proc 80, 3500–3506. doi: 10.1016/j.matpr.2021.07.281

Crossref Full Text | Google Scholar

Pandian, J. A., Kumar, V., Geman, O., Hnatiuc, M., Arif, M., and Kanchanadevi, K. (2022). Plant disease detection using deep convolutional neural network. Appl. Sci. 12:6982. doi: 10.3390/app12146982

Crossref Full Text | Google Scholar

Parez, S., Dilshad, N., Alanazi, T. M., and Weon Lee, J. (2023). Towards sustainable agricultural systems: a lightweight deep learning model for plant disease detection. Comput. Syst. Sci. Eng. 47, 515–536. doi: 10.32604/csse.2023.037992

Crossref Full Text | Google Scholar

Peyal, H. I., Nahiduzzaman, M., Pramanik, M. A. H., Syfullah, M. K., Shahriar, S. M., Sultana, A., et al. (2023). Plant disease classifier: detection of dual-crop diseases using lightweight 2D CNN architecture. IEEE Access 11, 110627–110643. doi: 10.1109/access.2023.3320686

Crossref Full Text | Google Scholar

Pramudhita, D. A., Azzahra, F., Arfat, I. K., Magdalena, R., and Saidah, S. (2023). Strawberry plant diseases classification using CNN based on MobileNetV3-large and EfficientNet-B0 architecture. J. Ilm. Tek. Elektro Komput. dan Inf. (JITEKI) 9, 522–534. doi: 10.26555/jiteki.v9i3.26341

Crossref Full Text | Google Scholar

Ramesh, S., Hebbar, R., Niveditha, M., Pooja, R., Shashank, N., and Vinod, P. V. (2018) Plant disease detection using machine learning, in Proc. 2018 International Conference on Design Innovations for 3Cs Compute Communicate Control (ICDI3C), Bangalore, India, pp. 41–45.

Google Scholar

Rashid, J., Khan, I., Ali, G., Almotiri, S. H., AlGhamdi, M. A., and Masood, K. (2021). Multi-level deep learning model for potato leaf disease recognition. Electronics 10:2064. doi: 10.3390/electronics10172064

Crossref Full Text | Google Scholar

Sabrol, H., and Satish, K. (2016), "Tomato plant disease classification in digital images using classification tree," in Proc. 2016 international conference on communication and signal processing (ICCSP), pp. 1242–1246, IEEE.

Google Scholar

Sahu, P., Chug, A., Singh, A. P., and Singh, D. (2023). Classification of crop leaf diseases using image to image translation with deep-dream. Multimed. Tools Appl. 82, 35585–35619. doi: 10.1007/s11042-023-14994-x

Crossref Full Text | Google Scholar

Saleem, M. H., Potgieter, J., and Arif, K. M. (2020). Plant disease classification: a comparative evaluation of convolutional neural networks and deep learning optimizers. Plan. Theory 9:1319. doi: 10.3390/plants9101319

PubMed Abstract | Crossref Full Text | Google Scholar

Singh, C., and Singh, A. (2025). Enhancing the performance of ResNet50V2 for Rice leaf disease detection using squeeze-and-excitation (SE) blocks. 2025 International Conference on Automation and Computation (AUTOCOM). IEEE.

Google Scholar

Singla, R. S., Gupta, A., Gupta, R., Tripathi, V., Naruka, M. S., Awasthi, S., et al. (2023). Plant disease classification using machine learning. in Proc. 2023 International Conference on Disruptive Technologies (ICDT), Greater Noida, India. 409–413.

Google Scholar

SlideShare (n.d.) Grain smut of sorghum [slide presentation] Available online at: https://www.slideshare.net/slideshow/grain-smut-of-sorghum/238691657

Google Scholar

Thakur, P. S., Chaturvedi, S., Khanna, P., Sheorey, T., and Ojha, A. (2023). Vision transformer meets convolutional neural network for plant disease classification. Ecol. Inform. 77:102245. doi: 10.1016/j.ecoinf.2023.102245

Crossref Full Text | Google Scholar

Thomkaew, J., and Intakosum, S. (2023). Plant species classification using leaf edge feature combination with morphological transformations and SIFT key point. J. Image Graph. 11, 91–97. doi: 10.18178/joig.11.1.91-97

Crossref Full Text | Google Scholar

Tuğrul, B., Elfatimi, E., and Eryiğit, R. (2022). Convolutional neural networks in detection of plant leaf diseases: a review. Agriculture 12:1192. doi: 10.3390/agriculture12081192

Crossref Full Text | Google Scholar

U.S. Department of Agriculture. (2020). United Nations declares 2020 the international year of plant health. Available online at: https://www.usda.gov/about-usda/news/press-releases/2020/01/27/united-nations-declares-2020-international-year-plant-health

Google Scholar

Unal, Y. (2025). Integrating CBAM and squeeze-and-excitation networks for accurate grapevine leaf disease diagnosis. Food Sci. Nutr. 13:e70377. doi: 10.1002/fsn3.70377

PubMed Abstract | Crossref Full Text | Google Scholar

Vamsidhar, V. E., Rani, P. J., and Babu, K. R. (2019). Plant disease identification and classification using image processing. Int. J. Eng. Adv. Technol. 8, 442–446.

Google Scholar

Vo, H., Quach, L., and Ngoc, H. T. (2023). Ensemble of deep learning models for multi-plant disease classification in smart farming. Int. J. Adv. Comput. Sci. Appl. 14, 1045–1054. doi: 10.14569/IJACSA.2023.01405108

Crossref Full Text | Google Scholar

Woo, S., Park, J., Lee, J. Y., and Kweon, I. S. (2018). Cbam: convolutional block attention module. Proc. Eur. Conf. Comput. Vis., 3–19. doi: 10.1007/978-3-030-01234-2_1

Crossref Full Text | Google Scholar

Xian, T., and Ngadiran, R. (2021). Plant diseases classification using machine learning. J. Phys. Conf. Ser. 1962:012024. doi: 10.1088/1742-6596/1962/1/012024

Crossref Full Text | Google Scholar

Xu, Y., Kong, S., Gao, Z., Chen, Q., Jiao, Y., and Li, C. (2022). HLNET model and application in crop leaf diseases identification. Sustain. For. 14:8915. doi: 10.3390/su14148915

Crossref Full Text | Google Scholar

Yao, J., Tran, S. N., Garg, S., and Sawyer, S. (2024). Deep learning for plant identification and disease classification from leaf images: multi-prediction approaches. ACM Comput. Surv. 56, 1–37. doi: 10.1145/3639816

Crossref Full Text | Google Scholar

Keywords: edge computing devices, deep learning, convolutional neural network, plant disease classification, attention mechanism

Citation: Karthikeyan S, Charan R, Narayanan S and Jani Anbarasi L (2025) Enhanced plant disease classification with attention-based convolutional neural network using squeeze and excitation mechanism. Front. Artif. Intell. 8:1640549. doi: 10.3389/frai.2025.1640549

Received: 03 June 2025; Accepted: 17 July 2025;
Published: 12 August 2025.

Edited by:

Ruopu Li, Southern Illinois University Carbondale, United States

Reviewed by:

André Ortoncelli, Federal Technological University of Paraná, Brazil
Yavuz Unal, Sinop University, Türkiye

Copyright © 2025 Karthikeyan, Charan, Narayanan and Jani Anbarasi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sathiya Narayanan, c2F0aGl5YW5hcmF5YW5hbi5zQHZpdC5hYy5pbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.