Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 27 October 2025

Sec. Sustainable and Intelligent Phytoprotection

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1644271

Cross-scale detection and cross-crop generalization verification of tomato diseases in complex agricultural environments

Jinghuan HuJinghuan Hu1Jinying Li*Jinying Li2*Heyang WangHeyang Wang1
  • 1College of Information and Technology, Jilin Agricultural University, Changchun, China
  • 2College of Horticulture, Jilin Agricultural University, Changchun, China

In order to overcome the key challenges associated with detecting tomato leaf disease in complex agricultural environments, such as leaf occlusion, variation in lesion size and light interference, this study presents a lightweight detection model called ToMASD. This model integrates multi-scale feature decoupling and an adaptive alignment mechanism. The model innovatively comprises a dual-branch adaptive alignment module (TAAM) that achieves cross-scale lesion semantic alignment via a dynamic feature pyramid, a local context-aware gated unit (Faster-GLUDet) that uses a spatial attention mechanism to suppress background noise interference, and a multi-scale decoupling detection head (MDH) that balances the detection accuracy of small and diffuse lesions. On a dataset containing six types of disease under various weather conditions, ToMASD achieves an average precision of 84.3%,.by a margin of 4.7% to 12.1% over thirteen mainstream models. The computational load is compressed to 7.1 GFLOPs. Through the introduction of a transfer learning paradigm, the pre-trained weights of the tomato disease detection model can be transferred to common bean and potato detection tasks. Through domain adaptation layers and adversarial feature decoupling strategies, the domain shift problem is overcome, achieving an average precision of 92.7% on the target crop test set. False detection rates in foggy and strong light conditions are controlled at 6.3% and 9.8%, respectively. This study achieves dual breakthroughs in terms of both high-precision detection in complex scenarios and the cross-crop generalization ability of lightweight models. It provides a new paradigm for universal agricultural disease monitoring systems that can be deployed at the edge.

1 Introduction

Global agricultural diseases cause over 220 billion US dollars in economic losses each year. Among these, the early detection of leaf diseases in crops is a key part of precision plant protection (Astani et al., 2022). As an important economic crop cultivated worldwide, the control of tomato diseases directly affects crop yield, quality, and agricultural economic benefits. Although deep learning-based detection methods have made significant progress in single-crop scenarios, models generally perform poorly in cross-crop scenarios due to domain specificity and the scarcity of labelled data. This limitation is particularly evident in developing countries with limited resources: small-scale farmers cannot afford the development costs of multi-crop-specific models, and agricultural expert systems lack the ability to generalise heterologous disease features. This results in delayed disease warnings and the waste of prevention and control resources. Therefore, innovating and applying leaf disease detection technology is important for ensuring food security, promoting sustainable agricultural development, and advancing smart agricultural systems.

In natural farmland environments, leaf disease detection often faces multiple challenges, such as leaves shading each other, disease patches having diverse shapes, light conditions fluctuating and diseases having weak early symptoms. Traditional image processing methods mostly rely on colour space segmentation combined with texture feature extraction for classification. For example, Javidan et al. used K-means clustering to segment diseased areas and applied morphological operations to optimise edge detection, achieving a 98.97% accuracy rate under controlled lighting conditions (Javidan et al., 2023). Similarly, Bhagat et al. used a network search-based SVM for classification and detection of plant leaf diseases (Bhagat et al., 2020), while Rodríguez et al. collected potato canopy images using a drone equipped with a multispectral sensor (Rodríguez et al., 2021). They combined vegetation indices and machine learning algorithms to achieve early detection and severity assessment of potato late blight. Furthermore, Saleem et al. designed a leaf segmentation process based on the ExG index and the region-growing method and combined the proportion of the diseased area to assess severity (Saleem et al., 2024). However, traditional methods face insufficient generalization ability in complex farmland scenarios due to their heavy reliance on manual feature design and experience-driven parameter tuning, and are difficult to adapt to the multi-scale disease representation requirements in dynamic field environments.

The advent of CNN has precipitated a paradigm shift in the realm of agricultural disease detection, with end-to-end architectures predicated on single-stage detectors becoming the prevailing paradigm. This is primarily attributable to the enhanced efficiency exhibited by these architectures. In the context of tomato leaf disease detection, researchers frequently employ a combination of deep learning models and conventional image processing techniques to enhance the precision of lesion localization. For instance, Barbedo proposed a threshold segmentation approach based on the HSV color space and morphological processing to extract lesion areas and verified the feasibility of combining traditional methods with CNN (Convolutional Neural Network) (Barbedo, 2018). Similarly, R. et al. embedded an attention mechanism in a pre-trained residual CNN, combined with multi-spectral data to enhance lesion feature expression, improving the discrimination of lesion features in complex environments (R. et al., 2020). Furthermore, Cong et al. developed a lightweight Mask R-CNN variant, optimizing lesion boundary localization through the integration of superpixel segmentation and edge detection algorithms (Cong et al., 2023). As proposed by Shin et al., a feature extraction and data augmentation strategy was proposed, combining a CNN with RGB images (Shin et al., 2021). This strategy achieved an average accuracy of 92.18% in the detection of strawberry leaf powdery mildew. Despite the efficacy of the aforementioned method in certain contexts, it remains confronted with numerous challenges in the context of natural farmlands. The distinguishing characteristics of disease spots are often obscured by leaf occlusion in complex backgrounds, leading to ambiguity in identification (Debnath et al., 2023)This study adopts EfficientNetV2B2 as the lightweight backbone network to achieve efficient and accurate disease identification. sing the DL approach, tomato leaf disease identification achieves nearly 100% accuracy on a test dataset. Additionally, the presence of similar diseases can result in confusion regarding texture, and the identification of early disease spots with low contrast can be challenging (Saleem et al., 2024). Furthermore, the method’s accuracy in distinguishing cross-diseases with similar symptoms is often limited.

Currently, the field of plant disease detection generally faces the bottleneck of model generalization caused by domain differences. Existing research is mostly limited to customized training for single-crop diseases and is difficult to effectively transfer to heterologous crops. The detection of tomato leaf diseases in real-world agricultural settings is hindered by several key challenges: leaf occlusion and overlap.Different diseases share visual characteristics, leading to misclassification. To address this issue, this study proposes a cross-crop transfer learning framework that breaks through the domain shift limitations of cross-species disease recognition by sharing low-level feature representations and domain adaptation optimization strategies. Specifically, a CNN backbone model is trained with a tomato leaf disease dataset, and the transfer learning framework freezes the shallow feature extraction layers to retain the common texture and morphological features of crops and adapt to the specific phenotypes of target crop diseases, combined with adversarial training to minimize the distribution differences between domains. This achievement provides a cross-crop transfer learning paradigm for building a universal plant disease intelligent monitoring system and promotes the large-scale application of precision plant protection technology.

The primary contributions of this paper are as follows:

1. In order to address the challenges posed by the attenuation of features in small target disease spots and the failure to detect early disease spots, a novel dual-branch adaptive alignment module has been designed. Through dynamic feature alignment and cross-scale feature interaction, it significantly improves the detection accuracy and robustness of tomato leaf diseases in complex agricultural environments.

2. The Faster-GLUDet feature enhancement unit was integrated, which employs partial convolution and local context-aware gating mechanisms. This enhancement to the model’s noise suppression capabilities is achieved while maintaining its lightweight nature.

3. The construction of a multi-scale decoupled detection head was undertaken. The model achieves balanced detection of cross-scale diseases and efficient distinction between small disease spots, spreading lesions, and mixed diseases through hierarchical feature fusion and Group Normalization optimization.

2 Materials and methods

2.1 Data processing

2.1.1 Data source

The tomato leaf disease dataset used in this study was sourced from the “Tomato Leaf Diseases Detect” standardized dataset released by the Roboflow open platform. It contains six typical disease categories (bacterial spot, early blight, late blight, leaf mold, target spot, and black spot) and healthy leaf samples, covering the early, middle, and late stages of disease development. In total, it includes 3,469 high-resolution RGB images. The six tomato leaf diseases of interest in this study are highly prevalent in major tomato-growing regions worldwide, causing yield losses of 20% to 65% (Lu et al., 2018; Panno et al., 2021). Table 1 provides a detailed breakdown of the final image distribution across all categories after augmentation and splitting. The common Bean Dataset was captured at the Guoxin Modern Agricultural Base in Changchun City, Jilin Province, and the public dataset Bean Disease Dataset. The Potato dataset is from the public Potato disease dataset which includes three categories: health, early disease and late disease (https://gitcode.com/open-source-toolkit/829ec). The inclusion of these diverse datasets from different crops is intended to rigorously validate the transferability of the features learned by our model from tomatoes to other commercially important crops.

Table 1
www.frontiersin.org

Table 1. Data distribution.

2.1.2 Data enhancement

To enhance the robustness of the model, the dataset was expanded to 7370 images using data augmentation techniques. In order to prevent the original image and enhanced image from appearing simultaneously in the training set and validation set, the original image is initially divided into a training set, validation set, and test set in a ratio of approximately 8:1:1. Subsequently, five techniques including horizontal flipping, vertical flipping, grayscale, contrast adjustment, and brightness adjustment were randomly applied to the original data to enhance the image and labels. The enhanced example image is shown in Figure 1, and the label distribution is shown in Table 1.

Figure 1
A collage showing six images of a single leaf with various transformations. (a) Original figure in color. (b) Horizontal flip. (c) Vertical flip. (d) Grayscale version. (e) Contrast adjustment. (f) Brightness adjustment. Each transformation alters the appearance of the leaf slightly.

Figure 1. Data enhancement methods for tomato leaf images: (a) original image; (b) horizontal flip; (c) vertical flip; (d) grayscale conversion; (e) contrast adjustment; (f) brightness adjustment.

In order to simulate the complex weather changes in the real tomato cultivation environment, this study adopts the RGB channel synthesis technology based on the atmospheric scattering model to generate enhanced images with controllable weather features on 50% of the typical samples in the training set. The synthesis formula is shown in Equation 1.

I(x)=J(x)·t(x)+A·(1t(x))(1)

In this study, x denotes the pixel coordinate, I(x) signifies the synthesized image, J(x) represents the original image, the transmittance map t(x) is constrained within the interval [0.2, 0.8] and controls the weather intensity gradient, and the atmospheric light value A restricts the amplitude of illumination attenuation. Figure 2 presents an image of medium-intensity synthetic weather.

Figure 2
Six images of a leaf illustrating different weather and lighting conditions. (a) Original image of a leaf. (b) Leaf obscured by rain. (c) Leaf partially obscured by fog. (d) Leaf affected by solar flare. (e) Leaf with overexposure effects. (f) Leaf with falling snowflakes in the foreground.

Figure 2. Randomly select a leaf image for weather synthesis: (a) original image; (b) rain simulation; (c) fog simulation; (d) solar flare simulation; (e) overexposure simulation; (f) snow simulation.

2.2 Methodology

2.2.1 Tomato multi-scenario adaptive scale detector

While general-purpose detectors like YOLOv11 have demonstrated strong performance on common datasets, their inherent architecture is not optimally designed for the unique challenges presented by complex agricultural environments, such as severe scale variation of lesions, leaf occlusions, and pervasive background noise. These limitations often lead to feature misalignment, reduced sensitivity to small objects, and compromised robustness under fluctuating lighting conditions.

To address these specific issues, we propose the Tomato Multi-scenario Adaptive Scale Detector (ToMASD), a novel lightweight architecture specifically engineered for high-precision disease detection in real-world field settings. The overarching design philosophy of ToMASD is to achieve an optimal balance between computational efficiency and detection accuracy by introducing three dedicated core modules that work in concert throughout the feature extraction and fusion pipeline.

As illustrated in Figure 3, the Two-branch Adaptive Alignment Module (TAAM) is integrated into the backbone network. Its purpose is to dynamically align and calibrate multi-scale features at the earliest stage, effectively mitigating the semantic misalignment between healthy and diseased tissue regions caused by occlusion and scale variance. The Faster-Gated Linear Unit (Faster-GLUDet) is embedded within the neck network. This module acts as an adaptive feature refiner, leveraging a gating mechanism to suppress irrelevant background noise. The Multi-scale Decoupling Head (MDH) is designed as the detection head. It replaces the conventional coupled head with a decoupled structure, allowing for independent optimization for classification and regression tasks at different feature scales. This synergistic design ensures that ToMASD is uniquely capable of handling the complexities of agricultural disease detection.

Figure 3
Flowchart of a neural network model architecture consisting of three main sections: Backbone, Neck, and Head. The Backbone includes modules like TAAM, SPPF, and C2PSA, processing an input image. The Neck involves operations like Upsample, Concat, and Faster-GLUDet, interacting with CBS. The Head concludes with the MDH module, outputting an image featuring a plant leaf. Arrows indicate the flow of data between components.

Figure 3. ToMASD model structure diagram.

2.2.2 Two-branch adaptive alignment module

The present study focuses on the problems of insufficient feature extraction of small-scale targets and redundant shallow computations in the YOLOv11 backbone for leaf lesion detection. It proposes a novel dual-branch adaptive alignment module, the Two-branch Adaptive Alignment Module (TAAM), as shown in Figure 4. The module under discussion achieves efficient computation through a multi-level feature sharing architecture. Firstly, the Pointwise Spatial Attention Stem (PSAStem) is utilised as the shared initial extraction layer, which pre-calibrates the input features through 1×1 pointwise convolution and an adaptive mechanism, thereby enabling the network to form dynamic spatial focusing capabilities at the input stage. Subsequently, the feature maps are processed through dual paths. The primary pathway integrates two C3k2 modules,each containing three standard 3×3 convolutions with 64 output channels, and a standard 3×3 convolution to preserve intricate features. The secondary pathway employs a 1×1 dimension-reducing convolution (reducing channels by a factor of 2) and subsequently connects to the optimised PSABlock. The initial feature extraction module combines pointwise convolution and spatial attention mechanisms, enabling the network to prioritise key regions in the input image and enhance the dynamic focusing ability on key spatial regions in the input stage while maintaining computational efficiency.

Figure 4
Diagram illustrating the TAAM model with pathways of blocks and connections. The flow begins with an input leading to P4AStem, followed by C3k2, Conv, PSABlock, AAM, and DWConv nodes. It shows connections for two configurations of C3k2 with c3k set to false and true, highlighting a split, bottleneck, and concatenate process within each section. The PSABlock is noted for its attention mechanism followed by dual convolution layers.

Figure 4. TAAM internal structure diagram.

After the dual-branch channels process the features in parallel, they are connected to the Adaptive Alignment Module (AAM), as shown in Figure 5. The input dual-channel features F1 and F2 are first compressed in the channel dimension through 1×1 convolution layers to obtain F1’ and F2’, and then adaptive alignment weights - Adaptive Align Weight are generated through the cross-branch feature interaction layer to dynamically balance the contribution of the two paths. They are input into a 3×3 dilated convolution to capture long-range context dependencies and generate dynamic path selection weights α through the Sigmoid activation function. After spatial alignment of the two paths, the key region responses are enhanced through element-wise multiplication, as shown in Equations 2 and 3, where σ represents the Sigmoid activation function and W_α is a learnable weight matrix. Finally, multi-scale feature complementarity is achieved through element-wise addition, and the results are merged and output.

Figure 5
Diagram of a computational process starting with two features. Each feature undergoes a 1x1 convolution, followed by concatenation. A 3x3 convolution and sigmoid function are applied, then split. Element-wise multiplication with parameters occurs, followed by element-wise summation. The process concludes with another 1x1 convolution. Key symbols indicate operations of summation and multiplication.

Figure 5. AAM structure diagram.

x1_weight=σ(Wα1F1')    α,σ[0,1](2)
x2_weight=σ(Wα2F2')    α,σ[0,1](3)

The innovation of this module can be attributed to the following: The following two mechanisms are introduced: 1) A dynamic weight adjustment mechanism is employed to optimise branch weights in real time based on the semantic distribution of the input features. This mechanism alleviates the spatial offset problem of heterogeneous features, such as misalignment interference between leaf lesions and healthy tissues. 2) Through the collaborative design of dilated convolution and gated attention, computational redundancy is reduced while local details and global pathological patterns are jointly modelled.

2.2.3 Faster-gated linear unit

The neck network of YOLOv11 employs depthwise separable convolution and channel pruning strategies, which have been shown to enhance the recall rate of small targets while concurrently reducing the model’s parameters. Tomato leaf diseases frequently manifest as minute spots, and the receptive field of the P5 layer in the feature pyramid is overly extensive, which may impede the learning of small target features. Secondly, when the brown necrotic spots of tomato late blight are similar in colour to the soil, the feature pyramid network may confuse the target with the background. In order to address these issues, we propose Faster-GLUDet, whose core lies in enhancing the model’s ability to extract disease features in complex backgrounds through a gating mechanism while maintaining model lightweight. The Faster-GLUDet module integrates FasterNetBlock and Convolutional GLU (Chen et al., 2023; Shi, 2023), as illustrated in Figure 6.

Figure 6
Diagram illustrating a neural network architecture. The input passes through a FasterNet Block leading to a Convolutional Gate Linear Unit. The FasterNet Block contains a series of operations: PConv 3x3, followed by two Conv 1x1 layers, with their outputs added together. The Convolutional Gate Linear Unit involves an input branching into two Linear operations, a DWConv 3x3, followed by an activation function, and a Linear operation, with outputs combined using addition operations.

Figure 6. Faster GLUDet module structure diagram.

The module’s primary function is the extraction of feature units through the utilisation of FasterNet Block, employing a 3x3 Partial Convolution to extract spatial features from a quarter of the input channels. This approach results in a 25% reduction in computational load when compared with traditional convolution, while retaining edge detail information. Subsequently, two 1x1 convolutions are connected to perform channel dimension reduction and feature fusion, thereby ensuring the effective preservation of multi-scale disease features. In order to enhance the feature interaction between modules, a dynamic gated feature enhancement unit, known as the Convolutional Gated Linear Unit (ConvGLU), is connected after the Faster Block. The model employs a dual-branch convolution to generate feature maps and gating signals, dynamically suppressing background noise and enhancing the response in the lesion area through element-wise multiplication. In the gating branch of the traditional gated linear unit (GLU), ConvGLU decomposes the standard 3x3 convolution into a cascade structure of depthwise convolution (DWConv) and pointwise convolution (PWConv), and combines a linear projection layer and GELU activation function to construct a lightweight feature enhancement path. A lightweight 3x3 depthwise convolution operation is introduced prior to the activation function in order to construct a gating channel attention mechanism based on neighbourhood features. The design converts global channel attention into local context-aware dynamic weight adjustment through the local receptive field characteristics of the convolution kernel. This retains the important information filtering ability in the channel dimension while significantly reducing computational complexity. ConvGLU employs convolution operations to capture local features in the image, thereby enhancing its efficacy in processing local information in comparison to traditional FFN. It is also capable of adaptively enhancing small target features.

The integration of Faster-GLUDet has been demonstrated to enhance the model’s feature extraction and expression capabilities to a considerable extent. The lightweight design of FasterBlock provides low-latency input for CGLU, while the gated weight generation module of CGLU further optimises multi-scale feature interaction, enabling the model to maintain lightweight while enhancing the diversity and hierarchy of features. This, in turn, helps better capture the details and context information of the target object. The system has been engineered to achieve dynamic regulation of feature channels, thereby further enhancing the semantic segmentation and spatial understanding capabilities of features. The combination of these two approaches has been shown to more effectively fuse multi-scale features and alleviate the problem of information loss, significantly improving the model’s performance in challenging plant disease recognition tasks.

2.2.4 Multi-scale decoupling head

In order to address the issue of uneven detection accuracy of traditional detection heads for small and large-scale lesions, the Multi-scale Decoupling Head (MDH) has been proposed, as illustrated in Figure 7. The core process is as follows: MDH receives three different-scale feature maps - P3, P4 and P5 - from the Feature Pyramid Network (FPN) in parallel, which respectively carry high-resolution details, medium-scale information and large receptive field context, thereby constructing a multi-scale perception foundation. The features of each scale first enter a unified feature enhancement pathway, which is composed of a series of grouped normalized convolutional modules: First, the channel dimension is adjusted and fused through a 1×1 Conv_GN, and then two 3×3 Conv_GN modules are continuously used to enhance the spatial feature expression. At the same time, group normalization is utilized to ensure the stability of the model under small-batch training. After feature enhancement, the network flow is completely decoupled into two independent branches dedicated to their respective functions: The classification branch precisely extracts features through two consecutive 1×1 convolutional layers and ultimately outputs a probability graph with a dimension of nc, accurately determining the category of the target within each anchor box; The regression branch adopts the same structure, but its output dimension is 4 × reg_max, which indicates that it uses an advanced distributed focus loss mechanism. By predicting the discrete distribution of bounding box coordinates, it greatly improves the accuracy of lesion location, where reg_max defines the flexible maximum value of the distribution. Ultimately, the outputs of the two branches are respectively normalized and integrated through a scale layer to generate the final detection results.

Figure 7
Flowchart showing a network architecture. Three inputs, P3, P4, and P5, connect to a series of boxes labeled “Conv_GN” with dimensions 1x1 and 3x3. These connect to output layers labeled “cls” and “4*reg_max” with dimensions 1x1, followed by outputs labeled “nc” and “scale.” The diagram illustrates a convolutional network sequence.

Figure 7. MDH detection head structure.

3 Analysis of experimental results

3.1 Experimental platform and parameter settings

In this study, the image input size is set to 640×640 pixels. To accelerate the convergence speed, the initial learning rate is set to 0.01, the stochastic gradient descent algorithm (SGD) is used for training, the weight decay coefficient is set to 0.0005, the momentum factor is set to 0.937, a total of 200 periods, and the size of the training batch is set to 32 times, and the workers are set to 12. All the experiments are performed on a Linux server. All the experiments are realized on a Linux server, and the specific configuration of the experimental environment is shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Experimental environment configuration.

3.2 Analysis and evaluation of the identification results

3.2.1 Evaluation index

In this paper, the metrics Precision, Recall, and mAP are utilised to evaluate the detection performance of the model. TP, FP, and FN represent the number of true positive, false positive, and false negative samples, respectively. C denotes the set of object categories, and |C| is the total number of categories. As shown in Equations 4, 5 and 6

Precision=TPTP+FP(4)
Recall=TPTP+FN(5)
mAP=1|C|cCAP(c)(6)

P denotes the proportion of correctly detected disease samples among all positive detections, thereby reflecting the model’s capacity to avoid false positives. R signifies the proportion of correctly detected disease samples among the actual existing disease samples, thus measuring the model’s ability to reduce false negatives. AP quantifies the detection performance of the model for a single disease category by calculating the area under the precision-recall curve. mAP is the average of APs for all categories, and a higher mAP indicates that the model’s detection effect on various disease categories is more balanced and accurate.

3.2.2 Comparative experiments of different models

To comprehensively evaluate the detection and generalization performance of the proposed ToMASD model, we conducted extensive comparative experiments with thirteen state-of-the-art object detection models on the same tomato leaf disease dataset. As summarized in Table 3, ToMASD achieved the highest scores in both precision and mAP, significantly outperforming all other contenders. RT-DETR achieves high accuracy but at the cost of high computational complexity, making it unsuitable for edge deployment. In comparison with the unimproved YOLOv11n, ToMASD has increased P, mAP, and Recall by 6.6%, 7.8%, and 5.9%, respectively, demonstrating its superior ability in target localization and classification in complex scenarios. Despite the fact that YOLOv11n exhibits a modestly diminished number of FLOPs in comparison with ToMASD, a notable deterioration in accuracy is evident, suggesting the potential for optimisation shortcomings within the feature ex-traction process. Despite the advantages in parameter quantity and computational cost of YOLOv5n and YOLOv11n, their accuracy still lags significantly behind ToMASD, further confirming the dual improvements in accuracy and recognition performance of the ToMASD model. Whilst sustaining its position of being lightweight, it has considerably surpassed the constraints of prevailing algorithms in the trade-off between ac-curacy and computational resource consumption. As illustrated in Figure 8, a performance comparison of the six models with the highest mAP is presented. A selection of six algorithms with comparable performance was made for the purpose of a comprehensive comparison, as illustrated in Figure 9. The performance of the metric is optimised by the distance of each axis of the curve from the intersection point. The area enclosed by the curve is positively correlated with the strength of the algorithm’s comprehensive performance. The comparison results demonstrate that the ToMASD model proposed in this paper exhibits advantages in all metrics, not only improving performance but also achieving lightweight, thus rendering it more suitable for practical scenarios.

Table 3
www.frontiersin.org

Table 3. Comparison of object detection results of different algorithms.

Figure 8
Bar chart comparing the performance of different models: YOLOv5n, YOLOv8n, YOLOv9s, YOLOv11n, YOLOv11s, RT-DETR, and ToMASD. Metrics shown are P%, mAP%, and FLOPs(G). All models show high P% and mAP% values between 60 and 100, while FLOPs(G) values vary significantly, with RT-DETR showing the highest value.

Figure 8. Performance bar charts of six models.

Figure 9
Radar chart comparing object detection models YOLOv5n, YOLOv8n, YOLOv9s, YOLOv11n, YOLOv11s, RT-DETR, and ToMASD across five metrics: Precision, mAP, Recall, Parameters, and FLOPs. Models are represented in different colors as indicated by the legend.

Figure 9. Comprehensive comparison of performance of six models.

To verify the performance of the TAAM module, this study systematically evaluated the effects of four mainstream attention modules as the Stem, as shown in Table 4. PSA achieved the best balance between accuracy and computational efficiency, reducing the computational complexity by 17.7% compared to the ECA module with the sec-ond-highest accuracy, verifying the performance of PSA.

Table 4
www.frontiersin.org

Table 4. Comparison of different attention modules.

Figure 10 shows the detection performance of the five models with the highest accuracy under different weather conditions, where gray boxes indicate missed detec-tions and black boxes indicate false positives. Under foggy conditions, except for To-MASD, the other four models misjudged fog points as diseases. When the light intensity was too high, the comparison models also failed to effectively suppress the exposed areas, resulting in missed detections of some real lesions. Figure 11 shows the detection results of the four lightweight models in conventional environments. Comprehensive comparative analysis shows that ToMASD exhibits the most superior performance under various complex weather conditions and has superior feature extraction and denoising capabilities in real agricultural environments. Figure 12 shows ToMASD’s detection of potato and common bean leaf diseases, and the experimental results show that ToMASD maintains high accuracy in the cross-crop task, achieving 92.1% and 93.5% accuracy in the detection of bean and potato leaf diseases, respectively, demonstrating its efficient generalization ability and transferability nature of cross-species training. Table 5 shows the recognition accuracy of ToMASD for different spots, Figure 13 shows the confusion matrix of this experiment, the model still maintains stable recognition performance in the category imbalance dataset, the recognition accuracy of Late Blight, Early Blight is close to 90%, Leaf Mold has a similar chromaticity of the yellow spot and the healthy tissues, which leads to the relatively low detection accuracy. Bacterial Spot and Target Spot have similar water-damaged spot characteristics, but the model still achieved 82.1% and 74.1% mAP values through multi-scale texture analysis, indicating the effectiveness of the feature decoupling mechanism.

Figure 10
Grid of images comparing various YOLO and RT-DETR object detection models. Each row represents a model version: YOLOv8s, YOLOv10n, YOLOv11s, RT-DETR, and ToMASD. Images include leaves with overlapping boxes in different colors indicating detection results, along with numerical confidence scores in various hues.

Figure 10. Detection results of five models under different weather conditions (gray boxes indicate missed detections and black boxes indicate false positives).

Figure 11
Grid of images comparing plant leaf detection across four models: YOLOv8n, YOLOv10n, YOLOv11n, and ToMASD. Each model produces different bounding boxes and confidence scores in four rows, indicating variations in object detection performance on the leaves.

Figure 11. Detection results of the four models in a conventional detection environment.

Figure 12
Images showing bean and potato leaf diseases detected by different machine learning models: YOLOv8n, YOLOv10n, YOLOv11n, and ToMASD. The bean images illustrate viral and rust infections, while the potato images display early and general late blight. Each row represents the same leaf, showcasing various detection confidence levels.

Figure 12. The detection results of leaf diseases of common beans and potatoes by ToMASD: (a) bean disease, (b) tomato disease. The models from left to right are YOLOv8n, YOLOv10n, YOLOv11n.

Table 5
www.frontiersin.org

Table 5. Evaluation indicators for different diseases.

Figure 13
Normalized confusion matrix displaying predicted versus true classifications. Diagonal values indicate prediction accuracy for each class, with the highest accuracy being 1.00 for class 2. Color intensity corresponds to prediction accuracy, with darker shades representing higher values.

Figure 13. Confusion matrix.

3.2.3 Ablation experiment

The proposed ToMASD model is based on YOLOv11n and has been optimised by introducing TAAM, Faster-GLUDet, and MDH. In order to evaluate the performance of each optimisation module, an experiment was conducted using the variable control method. The training and testing were carried out on the same dataset and training parameters, and the results are shown in Table 6. It is evident that Model C attained an mAP% of 77.9%, while concurrently sustaining a minimal computational cost. However, Model D, which introduced TAAM and Faster-GLUDet, exhibited an increase in computational cost to 12 FLOPs/G, attributable to parameter redundancy, yielding an accuracy of only 79.2%. The experiments indicated that the joint application of TAAM and MDH caused feature decoupling conflicts. The Model E achieved an 82.1% P% and 80.1% mAP%, thereby demonstrating the viability of multi-module collaborative optimisation through its lightweight design. The ToMASD model proposed in this paper was found to achieve the optimal balance in parameters, computational efficiency, and performance. In comparison with the baseline model A, it enhanced the accuracy by 9.4% whilst escalating the computational cost by a mere 0.4G. This outcome serves to demonstrate the efficacy of the collaborative design of multi-dimensional attention mechanisms and lightweight architectures for object detection tasks.

Table 6
www.frontiersin.org

Table 6. Results of model improvement ablation experiment.

3.2.4 Heat map visualization analysis

The present study employed the gradient-weighted class activation mapping technique (Selvaraju et al., 2017) to visualise the small target detection mechanism of the ToMASD model (Figure 14). The experimental findings demonstrate that the tomato leaf disease detection model, based on transfer learning, exhibits adapted feature capabilities in different crop disease recognition tasks. When the model is transferred from the source domain of tomatoes to the target domain of common beans, the heatmap analysis indicates that the lesion areas can still be effectively captured, although the the extent of the activated regions of highlighted areas is lower than that in the source domain. This suggests that the model has initially acquired the ability to locate disease spots across species through transfer learning. When the heatmap is transferred to potatoes with more distinct morphological features, more concentrated highlighted areas are shown, which may be related to the reticulate vein structure of potato leaves, thereby enhancing the distinguishability of texture features. It is noteworthy that the heatmaps of all three crops demonstrate a substantial contrast between the lesion areas and healthy tissues, thereby confirming that the model, while retaining key pathological features, has achieved adaptive adjustments to different crop leaf diseases through weight transfer. This cross-species disease recognition capability provides a feasible solution for intelligent diagnosis of multiple crop pests and diseases under resource-limited conditions.

Figure 14
Thermal images display leaves of tomato, bean, and potato plants analyzed for stress or disease detection. The images show varying color patterns, with highlighted sections indicating different temperature levels, marked by rectangles.

Figure 14. Feature visualization of leaf disease patterns across different crops: (a) tomato; (b) common bean; (c) potato.

4 Discussion

4.1 Resource identification initiative

The present study proposes a detection model, ToMASD, which integrates multi-scale feature fusion and dynamic attention mechanisms, with a view to addressing the challenges of tomato leaf disease detection in complex agricultural environments. The prevailing field of agricultural disease detection frequently grapples with challenges such as leaf occlusion, a wide spectrum of lesion morphologies, and intricate lighting conditions. Conventional methodologies are predicated on manually designed features, which are challenging to adapt to the multi-scale representation requirements of dynamic field environments. Despite its promising performance, ToMASD has several limitations: real-time deployment, Although lightweight, the model may still struggle on low-end edgedevices. Knowledge distillation or quantization could further compress the model. The research under discussion addresses the issue of misalignment between diseased and healthy tissues by designing a Two-branch Adaptive Alignment Module with a dynamic weight allocation mechanism. Secondly, the Faster-GLUDet module enhances noise suppression capabilities while maintaining a lightweight model through a local context-aware gating unit. The innovative architecture of the model integrates deep convolution and pointwise convolution to construct gating signals, thereby achieving a substantial reduction in background false detection rates, particularly in conditions characterized by foggy and strong light. It is evident that the multi-scale decoupled detection head (MDH) successfully achieves balanced detection of both small and diffuse lesions. This is achieved through the implementation of group normalisation and the establishment of independent classification and regression branches. A series of ablation experiments were conducted, the results of which demonstrate that MDH enhances the mAP value of imbalanced datasets by 7.8%, particularly with regard to the recognition accuracy of low-contrast diseases such as leaf mold.

Despite the fact that ToMASD demonstrates superiority over existing models in a number of metrics, there is still scope for enhancement. For instance, the mAP for Black Spot detection is only 76.1%, indicating that the ability to distinguish similar texture diseases needs further optimization. It is recommended that future research explore the potential of knowledge distillation techniques to further compress the model size. In addition, the combination of spatio-temporal features could enhance the prediction of disease spread dynamics. In conclusion, ToMASD has been demonstrated to provide an efficient and reliable solution for precise disease diagnosis in complex agricultural scenarios through the collaborative design of multi-dimensional attention mechanisms and lightweight architectures. This suggests that there is significant practical value and potential for promotion.

This study further validates the potential of ToMASD model for transfer learning in cross-crop disease detection. The weights of the tomato disease detection model were migrated to the potato and common bean tasks, and the cross-crop features were fused by domain adaptive layer state, and an adversarial feature decoupling strategy was used to suppress the inter-domain distribution bias. Experiments show that the migration model achieves 93.5% and 92.1% accuracy values on the potato and common bean test sets, respectively. The technical innovation is that by decoupling the cross-crop shared features and crop-specific features, the model overcomes the problem of confusing the small brown spots of bean rust with the background leaf vein texture while retaining the accuracy of tomato disease detection, supporting the joint monitoring of multi-crop diseases. The results show that the lightweight architecture and domain adaptive mechanism of ToMASD provide a cross-species generalization paradigm for building a general-purpose agricultural disease monitoring system, which significantly reduces the model development cost for multi-crop disease detection.

5 Conclusions

This study proposed ToMASD, a novel lightweight detection model, to address the critical challenges of tomato disease detection in complex agricultural environments. The core contributions of this work are threefold, each validated by extensive experimental results:

Firstly, to mitigate the feature attenuation of small lesions and the misalignment interference between diseased and healthy tissues, we designed the Two-branch Adaptive Alignment Module (TAAM). This module dynamically aligns cross-scale features, which was a key factor in achieving the overall mAP of 81.7%, a significant improvement over all baseline models.

Secondly, to enhance feature representation while suppressing complex background noise, we integrated the Faster-GLUDet feature enhancement unit. Its local context-aware gating mechanism effectively reduced false positives, as evidenced by the low misdetection rates of 6.3% in foggy and 9.8% in strong light conditions, while maintaining a low computational cost of only 7.1 GFLOPs.

Thirdly, to balance the detection accuracy between tiny spots and large lesions, we developed the Multi-scale Decoupling Head (MDH). By employing Group Normalization and independent task-specific branches, the MDH ensured balanced detection, which is reflected in the stable performance across all six disease categories, even under severe class imbalance.

Future research can be extended in three aspects: first, integrating hyperspectral imaging technology to enhance the characterisation of chromaticity gradient diseases; second, verifying the generalization of the model in economic crops such as chili peppers, grapes. based on the current migration learning framework; third, exploring the deployment scheme of edge computing to build a low-power field monitoring network in combination with LoRaWAN wireless transmission, to realize spatio-temporal prediction of disease spreading and precise prevention and control. The study is based on a lightweight architecture and a low-power edge computing deployment scheme. Through the deep integration of lightweight architecture and migration learning technology, this study provides a scalable solution for intelligent diagnosis of agricultural diseases, promotes the digital transformation of disease monitoring from single-crop scenarios to multi-species and multi-environment collaborative management, and assists the sustainable development of agriculture.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found here: https://gitcode.com/open-source-toolkit/829ec.

Author contributions

JH: Investigation, Writing – original draft, Formal Analysis, Software, Conceptualization, Data curation, Methodology, Writing – review & editing. JL: Conceptualization, Resources, Funding acquisition, Writing – review & editing, Writing – original draft, Visualization, Validation, Formal analysis. HW: Supervision, Data curation, Writing – review & editing, Validation, Writing – original draft, Resources.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This research was funded by the The Natural Science Foundation of Jilin Province (20240101209JC).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Astani, M., Hasheminejad, M., and Vaghefi, M. (2022). A diverse ensemble classifier for tomato disease recognition. Comput. Electron Agric. 198, 107054. doi: 10.1016/j.compag.2022.107054

Crossref Full Text | Google Scholar

Barbedo, J. G. A. (2018). Impact of dataset size and variety on the effectiveness of deep learning and transfer learning for plant disease classification. Comput. Electron Agric. 153, 46–53. doi: 10.1016/j.compag.2018.08.013

Crossref Full Text | Google Scholar

Bhagat, M., Kumar, D., Haque, I., Munda, H. S., and Bhagat, R. (2020). “Plant leaf disease classification using grid search based SVM,” in 2nd International Conference on Data, Engineering and Applications (IDEA). 1–6 (Bhopal, India: IEEE). doi: 10.1109/IDEA49133.2020.9170725

Crossref Full Text | Google Scholar

Chen, J., Kao, S., He, H., Zhuo, W., Wen, S., Lee, C.-H., et al. (2023). Run, don’t walk: chasing higher FLOPS for faster neural networks. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada doi: 10.1109/CVPR52729.2023.01157

Crossref Full Text | Google Scholar

Cong, P., Li, S., Zhou, J., Lv, K., and Feng, H. (2023). Research on instance segmentation algorithm of greenhouse sweet pepper detection based on improved mask RCNN. Agronomy 13, 196. doi: 10.3390/agronomy13010196

Crossref Full Text | Google Scholar

Debnath, A., Md., M., Raihan, M., Samrat, N., Alsulami, M. M., Masud, M., et al. (2023). A smartphone-based detection system for tomato leaf disease using efficientNetV2B2 and its explainability with artificial intelligence (AI). Sensors 23, 8685. doi: 10.3390/s23218685

PubMed Abstract | Crossref Full Text | Google Scholar

Javidan, S. M., Banakar, A., Vakilian, K. A., and Ampatzidis, Y. (2023). Diagnosis of grape leaf diseases using automatic K-means clustering and machine learning. Smart Agric. Technol. 3, 100081. doi: 10.1016/j.atech.2022.100081

Crossref Full Text | Google Scholar

Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., et al. (2022). YOLOv6: A single-stage object detection framework for industrial applications. doi: 10.48550/arXiv.2209.02976

Crossref Full Text | Google Scholar

Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., et al. (2016). SSD: single shot multiBox detector 21–37. doi: 10.1007/978-3-319-46448-0_2

Crossref Full Text | Google Scholar

Lu, J., Zhou, M., Gao, Y., and Jiang, H. (2018). Using hyperspectral imaging to discriminate yellow leaf curl disease in tomato leaves. Precis Agric. 19, 379–394. doi: 10.1007/s11119-017-9524-7

Crossref Full Text | Google Scholar

Panno, S., Davino, S., Caruso, A. G., Bertacca, S., Crnogorac, A., Mandić, A., et al. (2021). A review of the most common and economically important diseases that undermine the cultivation of tomato crop in the mediterranean basin. Agronomy 11, 2188. doi: 10.3390/agronomy11112188

Crossref Full Text | Google Scholar

R., K., M., H., Anand, S., Mathikshara, P., Johnson, A., and Menaka, R. (2020). Attention embedded residual CNN for disease detection in tomato leaves. Appl. Soft Comput. 86, 105933. doi: 10.1016/j.asoc.2019.105933

Crossref Full Text | Google Scholar

Redmon, J. and Farhadi, A. (2018). YOLOv3: an incremental improvement.

Google Scholar

Rodríguez, J., Lizarazo, I., Prieto, F., and Angulo-Morales, V. (2021). Assessment of potato late blight from UAV-based multispectral imagery. Comput. Electron Agric. 184, 106061. doi: 10.1016/j.compag.2021.106061

Crossref Full Text | Google Scholar

Saleem, S., Sharif, M. I., Sharif, M. I., Sajid, M. Z., and Marinello, F. (2024). Comparison of deep learning models for multi-crop leaf disease detection with enhanced vegetative feature isolation and definition of a new hybrid architecture. Agronomy 14, 2230. doi: 10.3390/agronomy14102230

Crossref Full Text | Google Scholar

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). “Grad-CAM: visual explanations from deep networks via gradient-based localization,” in 2017 IEEE International Conference on Computer Vision (ICCV). 618–626 (IEEE). doi: 10.1109/ICCV.2017.74

Crossref Full Text | Google Scholar

Shi, D. (2023). TransNeXt: robust foveal visual perception for vision transformers. doi: 10.1109/CVPR52733.2024.01683

Crossref Full Text | Google Scholar

Shin, J., Chang, Y. K., Heung, B., Nguyen-Quang, T., Price, G. W., and Al-Mallahi, A. (2021). A deep learning approach for RGB image-based powdery mildew disease detection on strawberry leaves. Comput. Electron Agric. 183, 106042. doi: 10.1016/j.compag.2021.106042

Crossref Full Text | Google Scholar

Wang, C.-Y., Bochkovskiy, A., and Liao, H.-Y. M. (2022). YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. doi: 10.1109/CVPR52729.2023.00721

Crossref Full Text | Google Scholar

Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., et al. (2024a). YOLOv10: real-time end-to-end object detection. ArXiv, abs/2405.14458. doi: 10.1007/978-3-031-72751-1_1

Crossref Full Text | Google Scholar

Wang, C.-Y., Yeh, I.-H., and Liao, H.-Y. M. (2024b). YOLOv9: learning what you want to learn using programmable gradient information.

Google Scholar

Keywords: tomato leaf disease, precision agriculture, agricultural artificial intelligence, multi-scale detection, transfer learning

Citation: Hu J, Li J and Wang H (2025) Cross-scale detection and cross-crop generalization verification of tomato diseases in complex agricultural environments. Front. Plant Sci. 16:1644271. doi: 10.3389/fpls.2025.1644271

Received: 10 June 2025; Accepted: 10 October 2025;
Published: 27 October 2025.

Edited by:

Chunlei Xia, Chinese Academy of Sciences (CAS), China

Reviewed by:

Anjan Debnath, Khulna University of Engineering & Technology, Bangladesh
Sreedevi Alampally, K L University, India

Copyright © 2025 Hu, Li and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jinying Li, bGlqaW55aW5nQGpsYXUuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.