- 1Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, India
- 2School of Computer Science & Engineering, Sandip University, Nashik, Maharashtra, India
- 3Department of Computer Science, College of Computer Science, Applied College Tanumah, King Khalid University, Abha, Saudi Arabia
- 4Department of Computer Science and IT, Government Sadiq College Women University, Bahawalpur, Pakistan
- 5Department of Information Systems, College of Computer and Information Sciences, Princess NourahbintAbdulrahman University, Riyadh, Saudi Arabia
- 6College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
Introduction: The rapid growth of the global population and intensive agricultural activities has posed serious environmental challenges. In response, there is an increasing demand for sustainable agricultural solutions that ensure efficient resource utilization while maintaining ecological balance. Among these, intercropping has gained prominence as a viable method, promoting enhanced land use efficiency and fostering environment for crop development. However, disease management in intercropping systems remains complex due to the potential for cross-infection and overlapping disease symptoms among crops. Early and precise illness recognition is, therefore, critical for sustaining crop condition and efficiency.
Methods: This study introduces an intelligent intercropping framework for early leaf disease detection, utilizing hyperspectral imaging and hybrid deep learning models for precision agriculture. Hyperspectral imaging captures intricate biochemical and structural variations in crops like maize, soybean, pea, and cucumber—subtle markers of disease that are otherwise imperceptible. These images enable accurate identification of diseases such as rust, leaf spot, and complex co-infections. To refine disease region segmentation and improve detection accuracy, the proposed model employs the synergistic swarm optimization (SSO) algorithm. A phase attention fusion network (PANet) is utilized for deep feature extraction, minimizing false detection rates. Furthermore, a dual-stage Kepler optimization (DSKO) algorithm addresses the challenge of high-dimensional data by choosing the most applicable landscapes. The disease classification is performed using a random deep convolutional neural network (R-DCNN).
Results and discussion: Experimental evaluations were conducted using publicly available hyperspectral datasets for maize–soybean and pea–cucumber intercropping systems. The suggested ideal attained remarkable organization accuracies of 99.676% and 99.538% for the respective intercropping systems, demonstrating its potential as a robust, non-invasive tool for smart, sustainable agriculture.
1 Introduction
Agriculture is the primary source of food, revenue, and jobs and contributes significantly to the global economy. Agriculture generates 18% of the country’s GDP and raises the employment rate to 53% in India and other low- and middle-income nations with large numbers of farmers. Because crop diseases drastically lower production, they have become a nightmare (Jamjoom et al., 2023). When dealing with several illnesses or a variety of planting circumstances, the conventional approaches are not very flexible (Zhao et al., 2024). Improved decision-making in agricultural production management is facilitated by the early detection of plant diseases. In addition to a back propagation neural network, conventional approaches like SVM and K-means clustering algorithms (Li et al., 2022) have been employed for plant disease detection. Cucumber leaf disease is classified using a two-stage model that combines DeepLabV3+ and U-Net in complicated backdrops (Raza et al., 2025). The Dice coefficient for lesion segmentation was 0.6914, the accuracy for illness classification was 92.85%, and the accuracy for leaf segmentation was 93.27%. Expanders and feature extraction are performed on point cloud data using the Generate Adversarial-Driven Cross-Aware Network (GACNet) (Yang et al., 2024). Through the dynamic combination of geographical location and feature attributes, GACNet improves the effectiveness of feature extraction. A cascaded incremental region network (Inc-RPN) (Hai et al., 2025) is used for accurate apple leaf disease detection in natural settings. A Coffee-Net model is used for accurate classification of coffee leaf diseases. Coffee-Net achieves 99.95% accuracy, outperforming ANN, Mask R-CNN, MobileNetV2, and ResNet50 by 0.6%, 4.32%, 0.02%, and 1.95% respectively (Zhang et al., 2023). The InceptionV3, MobileNetV1/V2, and VGG-16 models are optimized using pruning and quantization-aware training (Zhang et al., 2024). Although existing models demonstrate high accuracy in plant disease detection, they depend on large, labeled datasets and lack adaptability to new disease classes with limited data. To address these limits, this study presents a smart intercropping system by hyperspectral imaging and hybrid deep learning for accurate leaf disease detection and enhanced precision agriculture (Bidarakundi and Kumar, 2024; Da Silva and Almeida, 2024). Hyperspectral imaging captures intricate biochemical and structural variations in crops like maize-soybean and pea-cucumber, subtle markers of disease that are imperceptible. The key contributions of the planned work are given as trails.
1. The synergistic swarm optimization (SSO) algorithm is used to accurately segment the diseased regions from hyperspectral images of leaves. By focusing on the most relevant regions of infection, SSO enhances the spatial precision of the extracted disease areas, which improves the downstream classification accuracy.
2. To extract deep features from the segmented disease regions, phase attention fusion network (PANet) is integrated into the framework.
3. To tackle high-dimensionality issues inherent in hyperspectral data, the dual-stage Kepler optimization (DSKO) algorithm is used for feature optimization.
4. The random deep convolutional neural network (R-DCNN) model is used to classify leaf diseases within intercropping system, includes maize–soybean and pea–cucumber combinations which contributes to disease prediction even under complex conditions with co-occurring infections.
5. Experimental samples were collected on September 13, 2023, from maize–soybean and pea–cucumber intercropping fields in Fei Cheng City, Tai’an City, Shandong Province, China (Liu et al., 2024), ensuring the practical applicability of the framework.
The rest of this paper is organized as follows. Section 2 discusses the recent works on leaf disease prediction. The proposed methodology for smart intercropping system for accurate leaf disease detection is presented in Section 3. The consequences and conversation are presented in Section 4. The paper concludes in Section 5.
2 Related work
TinyResViT (Truong-Dang et al., 2025) is a lightweight efficient hybrid model that combines residual net (ResNet) and vision transformer (ViT) for leaf disease detection. The superfluous model weights are removed using the downsampling block that connects ViT and ResNet. With F1-scores of 97.92% and 99.11%, respectively, TinyResViT performs better on the plant village and Bangladeshi agricultural disease datasets. MobileH-Transformer (Thai and Le, 2025) combines a convolutional neural network (CNN) and Transformer for accurate leaf disease detection with minimal computation demands. The model obtains competitive F1-score values of 97.2% on the Maize leaf disease dataset and 96.80% on the Plant Village dataset, according to the results on publicly available datasets. A lightweight grape disease recognition method based on the GC-MobileNet model, for classification and fine-grained grading of diseases (Canghai et al., 2025). With an accuracy of 98.63 percent, GC-MobileNet outperforms MobileNetV3 by 6.51%. In grape vineyards, a real-time leaf detection system is used to identify and spray ill leaves, improving the efficacy of pesticide treatment (Khan et al., 2025). A deep learning algorithm can identify and classify six distinct diseases that affect potato leaves: nematodes, bacteria, viruses, fungi, phytophthora, and pests (Mhala et al., 2025). The LBPAttNet model integrates a lightweight coordinate attention mechanism into ResNet18 to enhance disease localization and reduce background interference (Wu P. et al., 2025).The model outperforms ResNet18 by 3.84% and 2.59%, respectively; with accuracy rates of 92.78% and 98.13%. Areas of interest in the crop leaf photos are found using an enhanced version of the U-Net segmentation algorithm (Chavan et al., 2025).The model’s F-measure was around 0.956 and its detection accuracy was better at 0. 982. For the accurate identification of grapevine leaf and fruit diseases, the ResNet50 model was improved using batch normalization (Sagar et al., 2025). During the validation stages, the model’s accuracy in distinguishing between healthy and sick grapevine leaves was 95%. LGENetB4CA combines modified EfficientNetB4 model with the LeafGabor filter (Van et al., 2025) which employs coordinate attention block to efficiently collect both spatial and channel-wise information.LGENetB4CA obtained an accuracy of 85.90% on COLD chili and 89.61% on JNUCLS for the COLD chili dataset. A semi-supervised method using modified pyramid scene parsing network (PSPNet) (Fan et al., 2025) for segmenting apple leaves. A fine-grained multi-label model based on transformers is used to categorize illnesses of apple leaves.
2.1 Problem description
Form related works, and numerous deep learning models (Liu et al., 2024; Dhanka and Maini, 2025; Han et al., 2025; Kamonsukyunyong et al., 2025; Logeswari et al., 2025; Mu et al., 2025; Patel, 2025; Qiao et al., 2025; Zhang C. et al., 2025; Zhang Z. et al., 2025) that have demonstrated promising results in the field of leaf ailment finding have been developed under monoculture conditions and exhibit limitations (Table 1) when applied to diversified cropping systems (Qin et al., 2024; Xu et al., 2024; Liang et al., 2025; Mohanty et al., 2025; Xu et al., 2025). TinyResViT, MobileH-Transformer, and GC-MobileNet have shown high accuracy on benchmark datasets; however, these methods rely heavily on ideal conditions and not generalize to real-world intercropping systems (Upadhyay et al., 2025). A critical observation from the literature review indicates that most existing works do not address leaf disease prediction in intercropping systems. To date, few studies have attempted to explore disease detection (Liu et al., 2024) within such systems, despite their increasing relevance in sustainable agriculture (Xu et al., 2023; Gong et al., 2024; Tu et al., 2024; Wang et al., 2024; Lu et al., 2025). Intercropping systems introduce unique challenges such as overlapping foliage, interspecies spectral interference, and a higher risk of co-infections—factors that are typically overlooked in conventional monoculture-based models (Liu et al., 2024; Qin et al., 2024; Xu et al., 2024; Cai et al., 2025; Chi et al., 2025; Christy and Jeyaraj, 2025; Dhanka and Maini, 2025; Han et al., 2025; Kamonsukyunyong et al., 2025; Liang et al., 2025; Li J. et al., 2025; Lingayya et al., 2025; Li Q. et al., 2025; Logeswari et al., 2025; Maranga et al., 2025; Mohanty et al., 2025; Mu et al., 2025; Patel, 2025; Qiao et al., 2025; Ratmele et al., 2025; Wang and Ruan, 2025; Wu M. et al., 2025; Xu et al., 2025; Zhang C. et al., 2025; Zhang Z. et al., 2025). To fill the research gaps, an intelligent framework for integrated early detection of foliar diseases is proposed, combining hyperspectral imaging with hybrid deep learning applied to precision agriculture. Hyperspectral imaging provides rich spectral information, capable of capturing subtle biochemical and structural changes in crops such as corn-soybean and pea-cucumber (Figure 1).

Table 1. Summary of research gaps from existing state-of-art works on leaf disease detection and classification.
3 Materials and methods
This study involves collecting hyperspectral images from Maize–soybean and pea–cucumber intercropping fields. Leaf samples, including both healthy and diseased ones, were captured under controlled lighting conditions using a hyperspectral imaging setup. The raw hyperspectral data underwent a series of preprocessing steps such as image calibration, noise removal, normalization, and enhancement to improve image quality. Leaf regions were segmented, and relevant features were extracted and refined. These processed features were then castoff to classify and classify various leaf ailments present in the intercropping systems. The overall architecture of the proposed smart intercropping system for leaf disease prediction is presented in Figure 2. Hyperspectral and leaf images are acquired from maize–soybean and pea–cucumber intercropping systems, where calibration, reference correction, noise removal, NDVI computation, leaf segmentation, and patch generation are applied to prepare the data. The acquired images undergo preprocessing steps such as spectral normalization, geometric correction, background subtraction, spectral smoothing, and image enhancement, ensuring uniform quality and reducing distortions. After preprocessing, the SSO algorithm segments the diseased regions from the healthy tissues, producing precise masks that highlight only the infected areas. These segmented regions are then processed through the phase attention fusion network (PANet) to extract discriminative deep features, while the dual-stage Kepler optimization (DSKO) algorithm reduces redundancy and optimizes the feature space. The refined features are classified using the random deep convolutional neural network (R-DCNN), which predicts crop-specific diseases. Within the maize–soybean intercropping system, the categories include normal, leaf spot, and rust, whereas in the pea–cucumber system, multiple diseases such as Ascochyta blight, powdery mildew, downy mildew, Fusarium wilt, cucumber spot, and anthracnose are identified. This integrated pipeline establishes a coherent flow from image acquisition and preprocessing to segmentation, feature extraction, optimization, and classification, thereby enabling accurate and scalable disease detection in diverse intercropping environments.
3.1 Data collection and preprocessing
The dataset used in this study is the publicly available hyperspectral dataset curated by Liu et al (Liu et al., 2024; Han et al., 2025; Zhang Z. et al., 2025). This dataset was collected from maize–soybean and pea–cucumber intercropping fields in Fei Cheng City, Tai’an, Shandong Province, China. It includes hyperspectral images of maize, soybean, pea, and cucumber leaves under various disease conditions. Specifically, the maize–soybean dataset contained healthy maize leaves, maize leaves with leaf spot, rust-infected maize leaves, and samples with combined infections, along with healthy and rust-infected soybean leaves. The pea–cucumber dataset (Figure 3) comprised healthy pea leaves as well as leaves infected with Ascochyta blight, powdery mildew, downy mildew, and Fusarium wilt. Cucumber samples included healthy leaves and those affected by angular leaf spot, powdery mildew, downy mildew, and anthracnose.

Figure 3. Sample leaf images from (A) Maize–soybean and (B) pea–cucumber intercropping fields with the disease class of corn normal, corn leaf spot, corn rust, corn hybrid, soybean normal, soybean rust” in Maize–soybean intercropping field; disease class of “pea normal, pea Ascochyta blight, pea powdery mildew, pea downy mildew, pea Fusarium wilt, cucumber normal, cucumber angular leaf spot, cucumber powdery mildew, cucumber downy mildew, and cucumber anthracnose” in pea–cucumber intercropping field.
● The ground-truth labeling of these datasets was performed by experienced plant pathologists as reported by Liu et al (Liu et al., 2024; Han et al., 2025; Zhang Z. et al., 2025), who applied phenotypic criteria such as lesion shape, size, color, and distribution on leaves for accurate annotation. In this study, we directly utilized these expert-verified labels for model training and evaluation.
● Details of hyperspectral imaging instrumentation and acquisition protocols are available in Liu et al (Liu et al., 2024; Han et al., 2025; Zhang Z. et al., 2025). In this work, we focused on preprocessing, model training, and evaluation using the curated dataset. To address potential class imbalance, patch generation was applied during preprocessing, ensuring sufficient representative samples for each disease category. The final dataset was split using stratified 10-fold cross-validation, which preserved the proportion of classes in training and testing sets while reducing bias in performance evaluation.
● During model training, hyperparameters such as learning rate, batch size, and number of epochs were tuned experimentally. A grid search strategy was employed, with the learning rate selected in the range 10-5 to 10-³, batch size varied between 16 and 64, and the epoch count adjusted to ensure convergence without overfitting. These design choices were based on preliminary trials and prior studies in hyperspectral plant disease detection.
● To illustrate the preprocessing workflow, Figure 4 presents representative raw images from the Liu et al. dataset alongside preprocessed images generated in this study, including patch extraction, normalization, and enhancement. Table 2 summarizes the key spectral bands identified in the dataset, highlighting differences between healthy and diseased leaves in the visible (450–700 nm) and near-infrared (700–740 nm) ranges, which correspond to physiological changes such as chlorophyll reduction, red edge shifts, and water stress. Table 3 reports the distribution of hyperspectral images across different crop and disease categories, with stratified splits used to ensure balanced representation in training, validation, and testing.

Figure 4. Example leaf images from the dataset: (A) raw input images of healthy and diseased leaves from maize–soybean and pea–cucumber intercropping systems; (B) corresponding preprocessed images after patch generation, normalization, and enhancement, illustrating the preparation of data for hyperspectral model training.

Table 2. Significant hyperspectral wavelengths distinguishing healthy and diseased leaves across maize–soybean and pea–cucumber intercropping systems, highlighting spectral variations associated with different plant diseases.
3.2 Target disease region segmentation
The target disease region segmentation plays a vital role by isolating only the infected portions of crop leaves from hyperspectral images while discarding irrelevant background and healthy regions. This step confirms that the subsequent stages of deep feature extraction and classification focus solely on the areas exhibiting disease symptoms, thereby improving accuracy and reducing false detections. To achieve this, a hybrid Synergistic Swarm Optimization (SSO) based on the Crowd Synchronization Algorithm (CSA) is employed (Logeswari et al., 2025). It is inspired by the collective behaviors of birds, fish, and ants, which rely on synchronization and cooperation to achieve optimal solutions. In this context, the input consists of hyperspectral images of maize, soybean, pea, and cucumber leaves, where each pixel carries intensity values across multiple spectral bands in the 400–1000 nm range. Unlike conventional segmentation methods such as K-means or Otsu’s thresholding, which fail under complex intercropping conditions, or deep segmentation models like U-Net that demand extensive annotated datasets and heavy computation, the proposed SSO-CSA hybrid dynamically adapts segmentation boundaries based on spectral similarities, swarm synchronization, and information sharing among candidate solutions (Zhang C. et al., 2025). This prevents premature convergence and ensures robust detection even in the presence of overlapping infections and multi-crop variations. The output of segmentation (Equation 1) process is a binary disease mask that highlights infected regions while suppressing background and healthy tissues, along with segmented hyperspectral sub-images that are passed to the PANet module for deep feature extraction. By focusing only on disease-affected regions, this segmentation approach reduces computational overhead and enhances the downstream classification accuracy of the system.
A matrix P represents fitness solution form the objective function (Equation 2).
where, un and ln stand for the vectors above and below each dimension of the issue space, respectively, and B implicitly represents the dimensions or variables Equation 3 of the given problem. Candidate solutions are updated with (P) (Dhanka and Maini, 2025).
where denotes the original optimal place g of the h-th applicant explanation, denotes g the current position of the h-th contender key, and g denotes the place of the h-th contender resolution value. The particles are based on the local and global attraction of states (Equation 4) (Kamonsukyunyong et al., 2025).
The value is the disinterest weight value (IWV), describes as trails (Equation 5):
When the inertia weight parameter (z), which dynamically regulates the ratio of exploration to exploitation (Equation 6), has an adaptive mechanism denoted by z. The following formula is used to get the personal best coefficient (PBC) (Patel, 2025):
where, solution number h is returned, Rand(xbest) is a unplanned resolution from the available entrant solutions, R1 is a random value, and Eps returns a tiny assessment. The global best coefficient (gbc) Equation 7 is calculated as follows (Qiao et al., 2025):
where, R2 is an accidental value representing the best comprehensive explanation and gives the solution number. The diversity preservation constant (dmc) is designed as trails (Equation 8) (Mu et al., 2025):
where, d2 is an extra hastening factor for the variety period and R5 is a random value. stands for the location in the cluster where the particle’s neighborhood diversity is maximized. The research gaps discussed in Table 1. Algorithm 1 describes the working process of target region segmentation using SSO.
3.3 Feature extraction
The feature extraction from the segmented target region of the disease transforms it into a set of meaningful descriptors that describe essential characteristics such as shape, texture, color, and structural patterns of the damaged area. In order for classification algorithms to accurately identify the type and severity of the disease, these extracted features are essential inputs. The fuzzy attention network (PANet) is used to efficiently extract features from the segmented disease areas in leaf images. PANet permits the network to focus more precisely on the most relevant input areas, such as diseased parts, while reducing the effect of healthy or non-relevant areas. Fuzzy attention components (Wang and Ruan, 2025) allow the model to selectively focus on exact feature maps that capture various visual cues such as texture, edges, and gradients at different levels of abstraction.
3.4 Feature optimization
In the context of identifying plant diseases, feature extraction often leads to a database that contains a variety of in-depth features. However, not all of these features are relevant or necessary for the classification of diseases. Feature optimization aims to reduce the dimensions of the dataset by identifying and retaining only the most important features that contribute efficiently to the classification task. By removing the less useful features, the model becomes more efficient, faster, and less prone to overfitting, thereby improving its performance and generalization capability. To optimize the features in this work, the two-stage Kepler optimization algorithm (DSKO) is used. DSKO is a metaheuristic algorithm inspired by nature and based on Kepler’s laws of planetary motion, which determine the movement of celestial bodies in space (Cai et al., 2025). The populace size is the numeral of planets that reflect the optimization problem’s decision parameters, is dispersed randomly over fuzzy sizes in the manner described in Equation 9:
where h-th represents candidate solution, is the numeral of applicant solution in the exploration space R1; and denotes the lesser and higher bounds of the g-th optimal parameter, separately. Where ai is the elliptical orbit semi-major axis at time s of object h, which is determined by Kepler’s third law as follows: the minimum value to prevent a divide-by-zero error. An absolute value randomly generated using a regular circulation to signify the orbital period of the objective. and ; denotes Euclidean distance normalization (Equation 10); defined as follows.
To inform the distance position of every objective from the Sun according to Equation 11 the previous steps:
where denotes an object H’s new position at time s + 1, denotes the object h’s current location at time s, VHS shows the velocity required for object h to move to the innovative situation, displays the optimal sun place, which is linked to the greatest explanation with the lowest fitness score, and F is shown as a flag to change the route of the exploration.The normalized values of and respectively (Equation 12),
where is the value of the fitness function with respect to each position of the object K at the current time s; Worst(t) denotes the solution candidate with the greatest fitness score (Equation 13). The term (), which denotes the normalized value of (), may be used to determine the Euclidian distance between and .
To accomplish exploration correctness, µ(s) is a purpose that exponentially debilities with time (s). The fitness optimization is expressed in Equation 14:
The procedure of feature optimization with DSKO is explicated in Algorithm 2.
3.5 Leaf disease prediction
Leaf disease prediction is process in the early detection and organization of plant diseases, based on the visual symptoms observed on the leaves of plants. In such an intercropping environment, the prediction of foliar diseases becomes more difficult due to the presence of various plants with different disease susceptibility profiles. Therefore, accurate disease prediction is crucial to prevent significant yield losses and ensure prompt intervention through appropriate disease management strategies. To achieve accurate and efficient predictions of leaf diseases in intercropping systems, a randomly used deep convolutional neural network (R-DCNN) is employed. R-DCNN is an advanced variant of the traditional deep convolutional neural network (DCNN), designed to enhance model performance by introducing random elements into the network structure and learning pathways. In R-DCNN, the convolution layer is used to learn parameters such as the weight matrix (n) and the dependence terms (E) of the convolution kernel (k). The complete convolution calculation is discussed in Equation 15 as follows (Li J. et al., 2025).
The following is the sample size of 1 × 2 if the pooling layer (d) employs uniform sampling (Equation 16).
For the pooling layer, s1 and s2 stands for the random input and output values, respectively. A feature network of R-DCNN is mapping of features from an input image to convolutional kernel and a processing function in convolutional layer (Equation 17).
To determine the neural output by using the nonlinear activation function is : The current layer additive bias k-th feature map is signified by E is the stimulation function (Equation 18) which usually starts at 0 (Chi et al., 2025).
The analysis’s findings demonstrate that the ReLU feature enhances the network’s capacity for recognition and learning. A fully connection layer is used to finalizes the random of nonlinear mapping and network size optimization (Equation 19).
The results of the ReLU feature improve the network’s recognition and learning capabilities. The objective function represents the combined weight and offset factors of layer 1 (Equation 20) input layer 2 units (Equation 21).
Forward propagation relies heavily on finding the appropriate intermediate stimulus value (Equations 22, 23) for each layer.
is the numeral of coatings in the neural net, h for the input layer, and for the output layer. The system’s main issue is the willpower of the excitation value at each buried layer in the neural system forward spread. Algorithm 3 describes how to use DSKO for leaf disease detection and categorization.
4 Results and discussion
This section presents the performance outcomes and comparative evaluation of the proposed model across multiple simulation scenarios. The evaluation focuses on: (i) quality of target region segmentation, (ii) effectiveness of deep feature extraction, (iii) impact of feature optimization, and (iv) comparison with existing state-of-the-art (SOTA) models for leaf disease prediction in maize–soybean and pea–cucumber intercropping systems. All experiments were conducted using the publicly available hyperspectral dataset curated by Liu et al. (2024), which contains samples from maize–soybean and pea–cucumber intercropping systems in China. This dataset was used for training, validation, and testing of the proposed framework. Table 4 summarizes the key hyperparameters and model configurations. The R-DCNN classifier was trained with a learning rate of 0.001, batch size of 32, and 100 epochs using the Adam optimizer. The model employed ReLU activation, a dropout rate of 0.3, a 3×3 kernel size, and five convolutional layers. The SSO algorithm used a swarm size of 30, 50 iterations, and exploration (α) and exploitation (β) factors of 0.6 and 0.4, respectively, initialized via a random Gaussian strategy with an accuracy-based fitness function. The DSKO module operated with a feature pool size of 200, a convergence tolerance of 1.00E–05, and dual-stage weighting factors of w1 = 0.7 and w2 = 0.3, ensuring optimized feature selection.
4.1 Results of segmentation algorithms
The results of the SSO algorithm was linked with present segmentation algorithms—K-means clustering (K-MC), fuzzy C-means (FCM), and particle swarm optimization (PSO)—using dice similarity coefficient and Jaccard index as evaluation metrics. Table 5 presents the comparative results of segmentation algorithms for leaf disease detection in intercropping systems. In the Corn-Soybean field, the SSO algorithm consistently outperformed K-MC, FCM, and PSO across all classes. Dice coefficient improvements ranged from 16.05% to 28.68%, while the Jaccard index increased by 29.74% to 49.93%. Similarly, in the Pea-Cucumber field, SSO achieved notable gains, with Dice increases up to 31.28% and Jaccard improvements reaching 53.75% over existing methods. These results highlight SSO’s superior segmentation accuracy for multi-crop disease detection. In the Pea-Cucumber field, SSO consistently outperformed K-MC, FCM, and PSO across all disease classes. Dice coefficient improvements ranged from 20.76% to 33.79% for pea diseases and 22.79% to 32.00% for cucumber diseases. Corresponding Jaccard index gains varied between 38.52% to 63.34% in pea and 41.30% to 55.44% in cucumber. These results highlight SSO’s superior segmentation performance, achieving 15% to over 60% improvements across metrics, ensuring highly accurate disease detection in intercropping.

Table 5. Results comparison of segmentation algorithms for intercropping based leaf disease detection.
4.2 Results analysis of prediction models
The exercise and challenging loss curves of the R-DCNN model for leaf disease prediction in Maize-soybean and pea-cucumber intercropping fields, as presented in Figure 5, show a consistent decline across increasing epochs, indicating effective model convergence and enhanced learning. In the Maize-soybean field, the exercise loss reduced from 0.402 at epoch 20 to 0.009 at epoch 1000, representing a 97.76% reduction. Similarly, the difficult loss concentrated from 0.417 to 0.032, yielding a 92.33% decrease, which reflects the model’s improved generalization ability. For the pea-cucumber intercropping field, the training loss dropped from 0.392 to 0.014, marking a 96.43% reduction, while the difficult loss decreased from 0.405 to 0.017, which corresponds to a 95.80% reduction. The exercise and difficult accuracy presentation of the R-DCNN model for leaf disease prediction, as illustrated in Figure 6, shows a substantial improvement across training epochs for both Maize-soybean and pea-cucumber intercropping fields.

Figure 5. Training and testing loss performance of R-DCNN for leaf disease prediction on (A) Maize-soybean and (B) pea-cucumber intercropping fields.

Figure 6. Training and testing accuracy performance of R-DCNN for leaf disease prediction on (A) Maize-soybean and (B) pea-cucumber intercropping fields.
The accuracy results of leaf disease detection in Maize-soybean intercropping fields, as shown in Figure 7, the performance comparison of models, PANet+R-DCNN and PANet+DSKO+R-DCNN, across various learning rates during the 10-fold validation process. The smallest improvement of 0.21% is observed at a learning rate of 0.002, where PANet+DSKO+R-DCNN achieved 0.967, compared to 0.959 for PANet+R-DCNN. PANet+DSKO+R-DCNN show an advantage, with an improvement in accuracy of approximately 0.56% across the ten-fold validation.

Figure 7. Accuracy results of leaf disease detection with varying learning rate of (A) PANet+R-DCNN and (B) PANet+DSKO+R-DCNN for Maize-soybean intercropping fields.
The correctness results of leaf illness discovery for pea-cucumber intercropping fields, presented in Figure 8, show the performance comparison between PANet+R-DCNN and PANet+DSKO+R-DCNN models across various learning rates. The results indicate that PANet+DSKO+R-DCNN consistently outperform PANet+R-DCNN, with improvement in accuracy. The ROC curves illustrated in Figure 9 shows the strong classification performance of the models across different crop disease classes. For both Maize-Soybean and Pea-Cucumber intercropping fields, the curves consistently stay well above the random guess line, indicating high true optimistic charges and low false confident charges.

Figure 8. Accuracy results of leaf disease detection with varying learning rate of (A) PANet+R-DCNN and (B) PANet+DSKO+R-DCNN for pea-cucumber intercropping fields.

Figure 9. ROC curves for disease classification performance in intercropping systems. (A) Maize-Soybean intercropping field with classes (B) Pea-Cucumber intercropping field with classes for PANet+DSKO+R-DCNN model.
Table 6 describes the performance evaluation of the proposed model for leaf disease detection in maize–soybean and pea–cucumber intercropping systems demonstrate consistently high accuracy across all disease classes. In the Maize–soybean field, normal maize leaves achieved an accuracy of 99.857%, with precision, sensitivity, specificity, and F-measure values all above 99.8%, indicating almost perfect classification. Leaf spot detection showed slight reductions in accuracy and specificity, while precision, sensitivity, and F-measure remained above 99.1%, highlighting minimal performance degradation. Rust-infected maize leaves achieved 99.364% accuracy, with all other metrics exceeding 99.2%, reflecting robust detection capabilities. Hybrid maize leaves demonstrated 99.396% accuracy, with minor variations in other metrics, showing that the model can accurately handle mixed infections. Soybean leaves, both normal and rust-infected, reached near-perfect scores, with normal leaves attaining 100% across all metrics and rust-infected leaves achieving 99.854% accuracy, with a slight decrease of 0.146% compared to normal leaves. In the Pea–cucumber field, normal pea leaves showed 99.413% accuracy, with corresponding high values for precision, sensitivity, specificity, and F-measure, reflecting consistent performance. Disease classes such as Ascochyta blight, powdery mildew, downy mildew, and Fusarium wilt exhibited accuracies of 99.132%, 99.053%, 99.024%, and 99.293%, respectively, representing minor reductions of 0.281%, 0.36%, 0.389%, and 0.12% compared to normal pea leaves, while maintaining high detection reliability. Similarly, cucumber leaves demonstrated strong performance, with normal leaves at 99.413% accuracy. Angular leaf spot, powdery mildew, downy mildew, and anthracnose achieved accuracies of 99.373%, 99.192%, 99.283%, and 99.182%, respectively, showing marginal decreases ranging from 0.041% to 0.221% compared to normal leaves. The results indicate that the model consistently delivers robust and reliable classification, with improvements ranging from 0.041% to 0.281% compared to the next best-performing disease classes, highlighting its effectiveness in real-world hyperspectral leaf disease detection scenarios.

Table 6. Performance metrics for leaf disease detection in maize–soybean and pea–cucumber intercropping systems, for each disease class.
4.3 Comparative analysis of proposed and SOTA models
Table 7 describes the recall of proposed PANet+R-DCNN and PANet+DSKO+R-DCNN models is compared with the existing SOTA models such as SVM+CARS, SVM+SPA, SVM+PCA, BiLSTM+CARS, BiLSTM+SPA, BiLSTM+PCA, DBO-BiLSTM+CARS, DBO-BiLSTM+SPA and DBO-BiLSTM+PCA. These results represent major improvements of up to 5.16% in Corn rust and 3.03% in Corn hybrid compared to even the top-performing DBO-BiLSTM models. PANet+R-DCNN yielded 98.568% recall in Corn rust and 98.148% in Corn hybrid, achieving over 3.86% and 1.54% gains respectively when compared to earlier baselines. Figure 10 describes the accuracy results comparison of proposed and SOTA models on Maize-soybean intercropping system. For the training phase, the proposed model PANet+DSKO+R-DCNN achieved the highest accuracy of 99.858%. Connected to SVM+CARS which recorded 93.6%, there is improvement of 6.66%. Over SVM+SPA and SVM+PCA with accuracy of 89.6% and 84.4% respectively, the gains are 10.26% and 15.26%. PANet+R-DCNN model recorded 98.748% the improvement is 1.05%, proving the effectiveness of the DSKO enhancement.
4.4 Impact of seasonal variations in disease detection for intercropping systems
Table 8 describes the seasonal variations in disease detection accuracy for Maize–soybean and pea–cucumber intercropping systems revealed significant differences, primarily influenced by temperature, lighting, and other environmental factors. During spring (Mar–May), the model achieved the highest detection accuracy, with values ranging from 94.034% to 97.52% for Maize–soybean and 94.014% to 97.411% for pea–cucumber. High-intensity illumination and crop stress caused performance to deteriorate in the summer, although autumn and winter showed comparatively poorer accuracy, especially in the winter when there was no natural light. Thus, variations in temperature and illumination have a significant impact on how accurately diseases are detected in these intercropping systems, with severe circumstances causing performance to noticeably deteriorate.

Table 8. Seasonal variations in disease detection accuracy (%) for Maize–soybean and pea–cucumber intercropping systems.
4.5 Statistics and comparative analysis
The descriptive statistics of recall performance for various models in the Maize-Soybean intercropping system, as shown in Table 9, reveal that the proposed PANet+DSKO+R-DCNN model achieved the highest mean recall of 99.916%, with minimal variability (Std. Dev. 0.148%) and a maximum recall of 100%, indicating highly consistent performance across all disease classes. The PANet+R-DCNN model also performed well with a mean recall of 99.453%, followed by DBO-BiLSTM+CARS with 98.55%. In contrast, classical models such as SVM+PCA and SVM+SPA showed lower mean recall of 81.467% and 87.883%, respectively, with much higher standard deviations, reflecting greater inconsistency and reduced reliability in detecting leaf diseases. The comparative analysis in Table 10 highlights the superiority of PANet+DSKO+R-DCNN over the state-of-the-art models. When compared to SVM+CARS, the proposed model showed the highest improvement in Corn hybrid by 25.54% and Corn rust by 23.16%, while Corn leaf spot and Soybean rust improved by 7.1% and 6.5%, respectively. There was no change in Corn normal and Soybean normal recall. Compared to SVM+SPA, the improvements were substantial, with Corn leaf spot increasing by 29.92%, Corn hybrid by 23.44%, and Soybean rust by 11.5%, while Corn rust improved by 6.55% and Corn normal remained unchanged. Against SVM+PCA, the proposed model achieved remarkable gains of 45.14% for Corn hybrid, 27.44% for Corn rust, 26.24% for Corn leaf spot, and 17.04% for Soybean rust, with minor improvement of 4.2% for Corn normal. Compared with DBO-BiLSTM+CARS, PANet+DSKO+R-DCNN showed moderate improvements, including 5.16% for Corn rust, 3.04% for Corn hybrid, and negligible decrease of 0.14% in Corn leaf spot. Similarly, against DBO-BiLSTM+SPA, the improvements were 9.64% in Corn hybrid, 7.85% in Corn rust, and 4.5% in Corn leaf spot, while the other classes remained unchanged. Compared to DBO-BiLSTM+PCA, the proposed model achieved 27.24% improvement in Corn hybrid, 5% in Corn leaf spot, 4.45% in Corn rust, and no change in other classes. The PANet+DSKO+R-DCNN model consistently outperformed both classical SVM-based models and deep learning-based SOTA models, demonstrating its robustness and effectiveness in accurately detecting leaf diseases across all classes in Maize-Soybean intercropping systems, with improvements ranging from minor gains of 4.2% to major increases of 45.14%.

Table 9. Descriptive statistics of recall performance (%) for various models in Maize-Soybean intercropping system.
5 Conclusion
This study presents an intelligent intercropping system that utilizes hyperspectral imaging and a hybrid deep learning framework for the uncovering of leaf ailments and the enhancement of precision agriculture. All experiments were conducted using the publicly available hyperspectral dataset curated by Liu et al. (2024) (Liu et al., 2024). The system uses the Synergistic Swarm Optimization (SSO) algorithm for precise segmentation of infested areas, the Phase Attention Network (PANet) for efficient feature extraction, and the Dual-stage Kepler Optimization (DSKO) algorithm for feature optimization. Subsequently, a random deep convolutional neural network (R-DCNN) is used to predict leaf diseases in both intercropping systems. Among the evaluated models, PANet+R-DCNN and PANet+DSKO+R-DCNN demonstrated exceptional performance, with PANet+DSKO+R-DCNN achieving the highest accuracy of 99.858% in exercise and 99.798% in testing, representing an improvement of 6.25% and 11.78% compared to the best traditional model, SVM+PCA. In terms of recall, the proposed model demonstrated significant improvements—up to 37.86% higher than SVM-based models and 9.09% better than BiLSTM-based models. These results confirm the reliability of the proposed models for precise and accurate leaf disease detection in intercropping systems.
5.1 Challenges and Limitations
Despite the promising performance of the proposed hyperspectral intercropping disease detection system, several challenges remain for practical deployment. The model’s accuracy is highly dependent on data quality, and field conditions such as motion artifacts, uneven illumination, and sensor noise can affect predictions. Environmental variability, including changes in lighting, temperature, and humidity, may further influence spectral measurements. Scalability is another concern, as real-time or large-scale applications require substantial computational resources and specialized hardware. Additionally, while the model performs well on maize–soybean and pea–cucumber systems, its generalization to other crops or regions needs further validation. Ensuring consistent hyperspectral imaging precision under field conditions also poses practical challenges. Addressing these limitations will be essential for robust and scalable deployment in precision agriculture.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author/s.
Author contributions
SG: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. VM: Conceptualization, Data curation, Formal analysis, Investigation, Resources, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. AR: Conceptualization, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing. MK: Conceptualization, Data curation, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. AI: Conceptualization, Formal analysis, Validation, Visualization, Writing – review & editing. BA: Funding acquisition, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. AA: Conceptualization, Funding acquisition, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research and/or publication of this article. Princess NourahbintAbdulrahman University Researchers Supporting Project number (PNURSP2025R440), Princess NourahbintAbdulrahman University, Riyadh, Saudi Arabia. The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/486/46.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Bidarakundi, P. M. and Kumar, B. M. (2024). Coffee-net: deep mobile patch generation network for coffee leaf disease classification. IEEE Sensors J. 25 (4), 7355–7362.
Cai, L., Zhao, S., Meng, F., and Zhang, T. (2025). Adaptive K-NN metric classification based on improved Kepler optimization algorithm. J. Supercomput 81, 1–44. doi: 10.1007/s11227-024-06559-y
Canghai, W., Xingxiang, G., Huanliang, X., and Huixin, H. (2025). Fine-grained recognition of grape leaf diseases based on transfer learning and convolutional block attention module. Appl. Soft Comput 172, 112896. doi: 10.1016/j.asoc.2025.112896
Chavan, P., Chavan, P. P., and Chavan, A. (2025). Hybrid architecture for crop detection and leaf disease detection with improved U-Net segmentation model and image processing. Crop Prot. 190, 107117. doi: 10.1016/j.cropro.2025.107117
Chi, P., Liang, R., Hao, C., Li, G., and Xin, M. (2025). Cable fault diagnosis with generalization capability using incremental learning and deep convolutional neural network. Electric Power Syst. Res. 241, 111304. doi: 10.1016/j.epsr.2024.111304
Christy, J. and Jeyaraj, P. R. (2025). Data-driven adaptive Lyapunov function based graphical deep convolutional neural network for smart grid congestion management. Electric Power Syst. Res. 238, 111163. doi: 10.1016/j.epsr.2024.111163
Da Silva, P. E. C. and Almeida, J. (2024). An edge computing-based solution for real-time leaf disease classification using thermal imaging. IEEE Geosci. Remote Sens. Lett. 22, 1–5.
Dhanka, S. and Maini, S. (2025). A hybrid machine learning approach using particle swarm optimization for cardiac arrhythmia classification. Int. J. Cardiol. 432, 133266. doi: 10.1016/j.ijcard.2025.133266
Fan, K. J., Liu, B. Y., Su, W. H., and Peng, Y. (2025). Semi-supervised deep learning framework based on modified pyramid scene parsing network for multi-label fine-grained classification and diagnosis of apple leaf diseases. Eng. Appl. Artif. Intell. 151, 110743. doi: 10.1016/j.engappai.2025.110743
Gong, L., Gao, B., Sun, Y., Zhang, W., Lin, G., Zhang, Z., et al. (2024). preciseSLAM: robust, real-time, liDAR–inertial–ultrasonic tightly-coupled SLAM with ultraprecise positioning for plant factories. IEEE Trans. Ind. Inf. 20, 8818–8827. doi: 10.1109/TII.2024.3361092
Hai, T., Shao, Y., Zhang, X., Yuan, G., Jia, R., Fu, Z., et al. (2025). An efficient model for leafy vegetable disease detection and segmentation based on few-shot learning framework and prototype attention mechanism. Plants 14 (5), 760. doi: 10.3390/plants14050760
Han, Y., Lv, M., Liu, J., He, S., Shi, W., Li, M., et al. (2025). Agronomic practices-driven response of nitrogen-related microorganisms. Plant Soil, 1–16. doi: 10.1007/s11104-025-07275-z
Jamjoom, M., Elhadad, A., Abulkasim, H., and Abbas, S. (2023). Plant leaf diseases classification using improved K-means clustering and SVM algorithm for segmentation. Computers Mater Continua 76 (1), 367–382. doi: 10.32604/cmc.2023.037310
Kamonsukyunyong, P., Katongtung, T., Srinophakun, T. R., and Sukpancharoen, S. (2025). Optimization of 3D printed drone performance using synergistic multi algorithms. Int. J. Thermofluids 26, 101058. doi: 10.1016/j.ijft.2025.101058
Khan, Z., Liu, H., Shen, Y., Yang, Z., Zhang, L., and Yang, F. (2025). Optimizing precision agriculture: A real-time detection approach for grape vineyard unhealthy leaves using deep learning improved YOLOv7 with feature extraction capabilities. Comput. Electron. Agric. 231, 109969. doi: 10.1016/j.compag.2025.109969
Li, J., Qiao, Y., Liu, S., Zhang, J., Yang, Z., and Wang, M. (2022). An improved YOLOv5-based vegetable disease detection method. Comput. Electron. Agric. 202, 107345. doi: 10.1016/j.compag.2022.107345
Li, J., Yang, Z., and Luo, Y. (2025). Intention inference for space targets using deep convolutional neural network. Adv. Space Res. 75, 2184–2200. doi: 10.1016/j.asr.2024.10.006
Li, Q., Wang, Z., Li, W., Hu, J., Rong, X., Bian, L., et al. (2025). Object segmentation of near surface magnetic field data based on deep convolutional neural networks. Comput. Geosci 196, 105847. doi: 10.1016/j.cageo.2024.105847
Liang, X., Wei, Z., and Chen, K. (2025). A method for segmentation and localization of tomato lateral pruning points in complex environments based on improved YOLOV5. Comput. Electron. Agric. 229, 109731. doi: 10.1016/j.compag.2024.109731
Lingayya, S., Kulkarni, P., Salins, R. D., Uppoor, S., and Gurudas, V. R. (2025). Detection and analysis of android malwares using hybrid dual Path bi-LSTM Kepler dynamic graph convolutional network. Int. J. Mach. Learn. Cybernet 16, 835–853. doi: 10.1007/s13042-024-02303-3
Liu, X., Meng, K., Zhang, K., Yang, W., Yang, J., Feng, L., et al. (2024). Discrimination of leaf diseases in maize/Soybean intercropping system based on hyperspectral imaging. Front. Plant Sci 15, 1434163. doi: 10.3389/fpls.2024.1434163
Logeswari, G., Thangaramya, K., Selvi, M., and Roselind, J. D. (2025). An improved synergistic dual-layer feature selection algorithm with two type classifier for efficient intrusion detection in IoT environment. Sci. Rep. 15, 8050. doi: 10.1038/s41598-025-91663-z
Lu, S., Zhu, G., Qiu, D., Li, R., Jiao, Y., Meng, G., et al. (2025). Optimizing irrigation in arid irrigated farmlands based on soil water movement processes: Knowledge from water isotope data. Geoderma 460, 117440. doi: 10.1016/j.geoderma.2025.117440
Maranga, J. O., Nnko, J. J., and Xiong, S. (2025). Learned active contours via transformer-based deep convolutional neural network using canny edge detection algorithm. Signal Image Video Process. 19, 222. doi: 10.1007/s11760-024-03795-w
Mhala, P., Bilandani, A., and Sharma, S. (2025). Enhancing crop productivity with fined-tuned deep convolution neural network for Potato leaf disease detection. Expert Syst. Appl. 267, 126066. doi: 10.1016/j.eswa.2024.126066
Mohanty, T., Pattanaik, P., Dash, S., Tripathy, H. P., and Holderbaum, W. (2025). Smart robotic system guided with YOLOv5 based machine learning framework for efficient herbicide usage in rice (Oryza sativa L.) under precision agriculture. Comput. Electron. Agric. 231, 110032. doi: 10.1016/j.compag.2025.110032
Mu, B., Li, Y., and Jiao, H. (2025). “Study on the synergistic construction of air-ground cooperative system and unmanned inspection system in the energy field,” in 2025 Asia-Europe Conference on Cybersecurity, Internet of Things and Soft Computing (CITSC). 391–396 (Rimini, Italy: IEEE).
Patel, N. (2025). Enhancing biofuel production with co-pyrolysis of distiller’s grains and waste polypropylene: synergistic effects and activation energy optimization with hybrid FLO-ENN approach. Chem. Eng. Processing-Process Intensificat 212, 110194. doi: 10.1016/j.cep.2025.110194
Qiao, S., Guo, Q., Shi, F., Wang, M., Zhu, H., Khan, F., et al. (2025). SIBW: A swarm intelligence-based network flow watermarking approach for privacy leakage detection in digital healthcare systems. IEEE J. Biomed. Health Inf. doi: 10.1109/JBHI.2025.3542561
Qin, W., Yang, X., Liu, C., and Zheng, W. (2024). A deep learning method based on YOLOv5 and SuperPoint-SuperGlue for digestive disease warning and cage location backtracking in stacked cage laying hen systems. Comput. Electron. Agric. 222, 108999. doi: 10.1016/j.compag.2024.108999
Ratmele, A., Dhanare, R., and Parte, S. (2025). Octave convolutional multi-head capsule nutcracker network with oppositional Kepler algorithm based spam email detection. Wireless Networks 31, 1625–1644. doi: 10.1007/s11276-024-03837-8
Raza, A., Pitafi, A. H., Shaikh, M. K., and Ahmed, K. (2025). Optimizing potato leaf disease recognition: insights DENSE-NET-121 and gaussian elimination filter fusion. Heliyon. 11 (3), e42318. doi: 10.1016/j.heliyon.2025.e42318
Sagar, N., Suresh, K. P., Sridhara, S., Patil, B., Archana, C. A., Sekar, Y. S., et al. (2025). Precision detection of grapevine downy and powdery mildew diseased leaves and fruits using enhanced ResNet50 with batch normalization. Comput. Electron. Agric. 232, 110144. doi: 10.1016/j.compag.2025.110144
Thai, H. T. and Le, K. H. (2025). MobileH-Transformer: Enabling real-time leaf disease detection using hybrid deep learning approach for smart agriculture. Crop Prot. 189, 107002. doi: 10.1016/j.cropro.2024.107002
Truong-Dang, V. L., Thai, H. T., and Le, K. H. (2025). TinyResViT: A lightweight hybrid deep learning model for on-device corn leaf disease detection. Internet Things. 30, 101495. doi: 10.1016/j.iot.2025.101495
Tu, B., Yang, X., He, B., Chen, Y., Li, J., and Plaza, A. (2024). Anomaly detection in hyperspectral images using adaptive graph frequency location. IEEE Trans. Neural Networks Learn. Syst. 36 (7), 12565–12579. doi: 10.1109/TNNLS.2024.3449573
Upadhyay, A., Chandel, N. S., Singh, K. P., Chakraborty, S. K., Nandede, B. M., and Kumar, M. (2025). Deep learning and computer vision in plant disease detection: a comprehensive review of techniques, models, and trends in precision agriculture. Artif. Intell. Rev. 58, 92. doi: 10.1007/s10462-024-11100-x
Van, H. T., Van Vu, G., Tuan, T. T., Vo, B., and Chung, Y. S. (2025). LGENetB4CA: A novel deep learning approach for chili germplasm Differentiation and leaf disease classification. Comput. Electron. Agric. 233, 110149. doi: 10.1016/j.compag.2025.110149
Wang, N., Wu, Q., Gui, Y., Hu, Q., and Li, W. (2024). Cross-modal segmentation network for winter wheat mapping in complex terrain using remote-sensing multi-temporal images and DEM data. Remote Sens. 16, 1775. doi: 10.3390/rs16101775
Wang, Y. and Ruan, Y. (2025). Improving the signal-to-noise ratio of single-channel earthquake data with an attention-based UNet3+ and phase-difference mask. IEEE Trans. Geosci. Remote Sens. 63, 1–10. doi: 10.1109/TGRS.2025.3555661
Wu, P., Liu, J., Jiang, M., Zhang, L., Ding, S., and Zhang, K. (2025). Tea leaf disease recognition using attention convolutional neural network and handcrafted features. Crop Prot. 122, 107118. doi: 10.1016/j.cropro.2025.107118
Wu, M., Zhang, X., Wang, Z., Tan, C., Wang, Y., and Wang, L. (2025). State of health estimation of lithium-ion batteries based on the Kepler optimization algorithm-multilayer-convolutional neural network. J. Energy Storage 122, 116644. doi: 10.1016/j.est.2025.116644
Xu, Y., Bai, Y., Fu, D., Cong, X., Jing, H., Liu, Z., et al. (2024). Multi-species weed detection and variable spraying system for farmland based on W-YOLOv5. Crop Prot. 182, 106720. doi: 10.1016/j.cropro.2024.106720
Xu, X., Fu, X., Zhao, H., Liu, M., Xu, A., and Ma, Y. (2023). Three-dimensional reconstruction and geometric morphology analysis of lunar small craters within the patrol range of the Yutu-2 Rover. Remote Sens. 15, 4251. doi: 10.3390/rs15174251
Xu, X., Zhou, B., Li, W., and Wang, F. (2025). A method for detecting persimmon leaf diseases using the lightweight YOLOv5 model. Expert Syst. Appl. 285, 127567. doi: 10.1016/j.eswa.2025.127567
Yang, T., Wei, J., Xiao, Y., Wang, S., Tan, J., Niu, Y., et al. (2024). LT-DeepLab: an improved DeepLabV3+ cross-scale segmentation algorithm for Zanthoxylumbungeanum Maxim leaf-trunk diseases in real-world environments. Front. Plant Sci 15, 1423238. doi: 10.3389/fpls.2024.1423238
Zhang, C., Cheng, J., Guo, Z., and Li, J. (2025). Unraveling surface roughness variations in SLM-GH4169 alloy polishing: A synergistic approach combining mechanistic modeling and machine learning algorithms. Mater Today Commun., 112505. doi: 10.1016/j.mtcomm.2025.112505
Zhang, W., Li, Z., Li, G., Zhuang, P., Hou, G., Zhang, Q., et al. (2023). GACNet: Generate adversarial-driven cross-aware network for hyperspectral wheat variety identification. IEEE Trans. Geosci. Remote Sens. 62, 1–14.
Zhang, Z., Yang, W., and Dong, Y. (2025). Faba bean–wheat intercropping reconstructed the microbial community structure in the rhizosphere soil of faba bean under F. commune and benzoic acid stress to alleviate Fusarium wilt in faba bean. Plant Soil 506 (1), 191–208. doi: 10.1007/s11104-023-06393-w
Zhang, H., Yang, J., Lv, C., Wei, X., Han, H., and Liu, B. (2024). Incremental RPN: hierarchical region proposal network for apple leaf disease detection in natural environments. IEEE/ACM Trans. Comput. Biol. Bioinf. 21 (6), 2418–2431. doi: 10.1109/TCBB.2024.3469178
Keywords: intercropping system, maize–soybean, pea–cucumber, hyperspectral imaging, deep learning, precisionagriculture
Citation: Goyal SB, Malik V, Rajawat AS, Khan M, Ikram A, Alabdullah B and Almjally A (2025) Smart intercropping system to detect leaf disease using hyperspectral imaging and hybrid deep learning for precision agriculture. Front. Plant Sci. 16:1662251. doi: 10.3389/fpls.2025.1662251
Received: 23 July 2025; Accepted: 16 September 2025;
Published: 07 October 2025.
Edited by:
Chengcheng Chen, Shenyang Aerospace University, ChinaReviewed by:
Yasin Kaya, Adana Science and Technology University, TürkiyeTri Handhika, Universitas Gunadarma Pusat Studi Komputasi Matematika, Indonesia
Abhishek Upadhyay, Central Institute of Agricultural Engineering (ICAR), India
Copyright © 2025 Goyal, Malik, Rajawat, Khan, Ikram, Alabdullah and Almjally. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Amna Ikram, YW1uYWlrcmFtQGdzY3d1LmVkdS5waw==; Bayan Alabdullah, YmlhbGFiZHVsbGFoQHBudS5lZHUuc2E=