Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 09 May 2023
Sec. Technical Advances in Plant Science

LightMixer: A novel lightweight convolutional neural network for tomato disease detection

Yi ZhongYi Zhong1Zihan TengZihan Teng2Mengjun Tong*Mengjun Tong1*
  • 1College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou, China
  • 2School of Design, The Hong Kong Polytechnic University, Hong Kong, Hong Kong, SAR, China

Tomatoes are among the very important crops grown worldwide. However, tomato diseases can harm the health of tomato plants during growth and reduce tomato yields over large areas. The development of computer vision technology offers the prospect of solving this problem. However, traditional deep learning algorithms require a high computational cost and several parameters. Therefore, a lightweight tomato leaf disease identification model called LightMixer was designed in this study. The LightMixer model comprises a depth convolution with a Phish module and a light residual module. Depth convolution with the Phish module represents a lightweight convolution module designed to splice nonlinear activation functions with depth convolution as the backbone; it also focuses on lightweight convolutional feature extraction to facilitate deep feature fusion. The light residual module was built based on lightweight residual blocks to accelerate the computational efficiency of the entire network architecture and reduce the information loss of disease features. Experimental results show that the proposed LightMixer model achieved 99.3% accuracy on public datasets while requiring only 1.5 M parameters, an improvement over other classical convolutional neural network and lightweight models, and can be used for automatic tomato leaf disease identification on mobile devices.

1 Introduction

Tomatoes are one of the most widely grown and consumed crops worldwide and are a high source of income for many agricultural countries (Al‐Gaashani et al., 2022). According to the United Nations Food and Agriculture Organization, the global production of tomatoes alone is expected to exceed 180 million tons by 2020 (Nations, 2020). China is currently ranked first among tomato-producing countries (Faostat, 2022). However, tomatoes are susceptible to various diseases during cultivation, which can significantly affect their yield (Xu et al., 2022). The early identification and detection of diseases can reduce the infection and spread of tomato crops and reduce the use of unnecessary agrochemicals. Early disease characteristics are expressed on plant leaves, and farmers or pathologists with extensive experience in tomato cultivation can effectively diagnose the type of crop disease and take necessary measures through visual inspection of various infected leaves (Cap et al., 2021). Although such traditional methods of disease inspection have a positive effect on the prevention of tomato diseases, they are only usable by professionals or empiricists, especially in developing countries where farmers in many regions are often unable to make effective diagnoses of plant diseases. In addition, subjective factors in personal perception, inefficiency, and labor costs are also issues that should be considered (Abbas et al., 2021). Therefore, the design of a simple, efficient, and accurate automatic tomato leaf disease diagnosis system that requires few computational parameters is of great importance to help farmers detect tomato diseases and improve tomato production. A lightweight model called LightMixer is proposed in this study. This model uses lightweight convolution to extract disease features, passes them to the deep network, and introduces the designed residual blocks to further reduce the number of parameters of the network and accelerate the computational efficiency of the model. The contributions of this study are as follows:

1. A novel lightweight deep learning model called LightMixer was proposed for the automatic detection of tomato leaf diseases under field conditions.

2. The depth convolution with Phish (DCWP) and light residual (LR) modules were carefully designed to improve deep feature fusion while reducing the number of model parameters.

3. Phish activation function was used to replace the commonly used rectified linear unit function to reduce the information loss of tomato leaf disease features and improve the nonlinear expression of the model.

4. Compared with other classical convolutional neural network (CNN) models, the proposed model achieved minimum memory cost and highest accuracy, making it more suitable for deployment in mobile devices or small embedded devices.

2 Related work

In recent years, with the development of machine learning and deep learning, the extensive use of computer vision techniques in agriculture has achieved remarkable results, particularly in the field of plant disease identification. Bhatia et al. (2020) proposed a hybrid support vector machine and logistic regression algorithm to extract the disease features of tomato powdery mildew to diagnose tomato diseases, achieving a 92.37% identification accuracy. Bhagat et al. (2020) used a grid search technique to optimize the support vector machine algorithm to classify plant leaf diseases. Chopda et al. (2018) built an Android application based on a decision-tree algorithm to effectively solve the classification problem of cotton crop diseases for local farmers. However, traditional machine learning image processing algorithms require the manual construction of disease features, which increases labor costs and cannot be adapted to identify multiple types of plant diseases. Therefore, deep learning algorithms that do not require manual feature construction are increasingly used to identify and diagnose plant diseases. Darwish et al. (2020) achieved a 98.2% recognition accuracy on a dataset containing five types of maize leaf images based on a combination of an integrated VGGNet architecture and an orthogonal particle swarm optimization algorithm to overcome the limitations in sample size and diversity. Fan et al. (2022) investigated migration learning and feature fusion to recognize plant leaves; the experimental results showed that the accuracy of the method exceeded 97% on a publicly available dataset. Chen et al. (2020) proposed the INC-VGGN network architecture for rice disease recognition by fusing the VGGNet architecture with the Inception module and achieved good recognition accuracy. Karthik et al. (2020) designed an attention-based deep residual network to identify tomato diseases; experimental results showed that the designed network architecture was able to identify the disease types with 98% accuracy. Zhao et al. (2021) improved the ResNet network architecture based on the attention module to extract various disease features of tomato leaves, achieving an average recognition accuracy of 96.81%. Lee et al. (2020) explored the identification of plant diseases using a recurrent neural network incorporating an attention mechanism, which has better generalization over public datasets compared to the classical CNN approach. L Lakshmanarao et al. (2021) applied the Convnets network architecture to identify nine tomato diseases. Agarwal et al. (2020) pre-trained the VGG16 network architecture, compressed the selected model, and validated it on a public dataset; experimental results showed that their proposed network architecture achieved an accuracy of 98%. Liu et al. (2022) chose the ResNext50 network architecture as the backbone and introduced expanded convolution and CA attention modules to design a DCCAM-MRNet network to identify six tomato leaf diseases. The experimental results showed that the classification accuracy of DCCAM-MRNet reached 94.3% with fewer parameters than the ResNext50 network architecture, which has a notable performance advantage.

Although the aforementioned studies strongly demonstrate the effectiveness of CNN architectures in the field of plant disease identification, these architectures are inevitably problematic because of their large number of parameters and high computational complexity. Therefore, designs focusing on lightweight CNNs have been widely proposed. Kamal et al. (2019) constructed lightweight CNNs for plant disease recognition on public datasets based on deep separable convolution using the MobileNet architecture, which has fewer parameters compared to other classical CNNs. Chen et al. (2022) replaced the standard convolutional approach in the DenseNet architecture, and Singh et al. (2021) verified the effectiveness of their lightweight model by fine-tuning several pre-trained models and comparing them with other classical CNNs. Barman et al. (2020) proposed the use of the MobileNet and SSCNN architectures to detect citrus leaf diseases in stages to achieve the efficiency of lightweight models. The experimental results showed that the algorithm had lower computational complexity and achieved a high accuracy of 99%. With similar accuracy, Bi et al. (2022) demonstrated a lower computational complexity using a MobileNet network architecture with deeply separable convolutions to identify two classes of apple diseases. The use of superior performance loss functions to optimize network architectures has also been proposed for the development of lightweight deep learning network architectures. For example, Zeng et al. (2022) used a new loss function to optimize the proposed lightweight LDSNet network architecture to recognize corn disease images based on a public dataset. The experimental results showed that the optimized network architecture achieved an accuracy of 95.4%, outperforming other classical lightweight models. Chen et al. (2021) used an enhanced loss function to optimize the MobileNetV2 model to identify rice diseases in complex backgrounds; they achieved an average accuracy of 98.48%, outperforming other classical methods. The aforementioned studies performed well in the field of plant disease identification, but the lightness of the model and accuracy of identification can be further explored.

3 Materials and methods

3.1 Image dataset acquisition and pre-processing

The data in this study were obtained from the PlantVillage public dataset (Arun Pandian and Gopal, 2019), which has 18835 tomato leaf images classified into ten categories, one healthy category and nine categories corresponding to different tomato leaf diseases. Table 1 shows the details of the dataset. Supplementary Figure 1 shows sample images for each tomato leaf category and describes the relevant information in the dataset, including the category labels, names of the diseases, and number of images used in the study for each category. To validate the generalization of the deep learning model, the image dataset in this study was randomly divided into an 80% training set, 10% validation set, and 10% test set.

TABLE 1
www.frontiersin.org

Table 1 Details of the dataset.

For a deep learning model, the more training data input into the model, the better the model learns. Therefore, several data enhancement methods were used to perform data enhancement on the training dataset before the deep learning model started training, including random image cropping, horizontal random flip, vertical random flip, and random rotation. All images were resized to 224 × 224 pixels. Supplementary Figure 2 shows the results obtained after applying the data enhancement methods to the tomato leaf images in the training set.

3.2 LightMixer model

In this study, a lightweight CNN model called LightMixer is proposed for identifying tomato leaf diseases in complex environments. The proposed LightMixer model focuses on both lower memory cost and higher accuracy to address the migration difficulties for mobile or small embedded devices. By designing and assembling the DCWP and LR modules, LightMixer can accurately extract features from tomato leaf disease images with complex backgrounds and achieve high accuracy with low-memory cost. LightMixer is a lightweight CNN model that is suitable for mobile and cross-connected wireless communication devices. The LightMixer structure is shown in Figure 1. It mainly consists of a convolutional block, a DCWP module, an LR module, a mean pooling layer, a Flatten layer, and a fully connected layer. A 3 × 3 convolutional layer is used in the first convolutional block of LightMixer to obtain a rich feature representation of the image. After the first convolution block, the DCWP module is connected to extract the deep feature information of the feature map without parameter redundancy. The LR module is connected after the DCWP module to increase the depth of the CNN with a small number of parameters to improve the nonlinear representation of the model while reducing the number of network parameters. The average pooling layer is used to reduce the spatial dimensionality of the feature map and increase the perceptual field of the model. The last classifier block uses the softmax function to implement disease classification. The configuration of the LightMixer model is listed in Table 2.

FIGURE 1
www.frontiersin.org

Figure 1 The structure of the LightMixer model.

TABLE 2
www.frontiersin.org

Table 2 Architecture of the LightMixer model.

3.3 Phish activation function

The activation function is a key component for minimizing the loss in deep neural networks and can both enhance the nonlinear representation and accelerate the computational efficiency of the entire model, thus improving the accuracy of tomato leaf disease identification. Compared with traditional activation functions introduced in CNN models, the Phish activation function outperforms most traditional activation functions in terms of classification; therefore, it is introduced in this study (Naveen, 2022). The Phish activation function consists of the existing GELU function (Hendrycks and Gimpel, 2016) with the Tanh function, and is defined as follows:

Phish(x)=xTanh(GELU(x))(1)

where x denotes the input, and the mathematical operations of the Tanh and GELU functions are shown in Equations (2) and (3), respectively.

Tanh(x)=exexex+ex(2)
GELU(x)=π2x(1+tanh1[2π(x+0.044715x3)])(3)

Therefore, by combining Equations (1), (2), and (3), the equation for the Phish function can be expressed as:

Phish(x)=xeπ2x(1+tanh1[2π(x+0.044715x3)])e π2x(1+tanh1[2π(x+0.044715x3)])e π2x(1+tanh1[2π(x+0.044715x3)])+eπ2x(1+tanh1[2π(x+0.044715x3)])(4)

where x denotes the input and Phish(x) denotes the nonlinear output after mathematical operations.

3.4 Depth convolution with phish

Depth-separable convolution reduces the number of parameters required for convolution computation by splitting the correlation between the spatial and channel dimensions; it has also been shown in some studies to improve the efficiency of convolution kernel parameters. As shown in Figure 2A, the standard convolution of the k × k kernel increases the number of channels in the feature map. Depthwise convolution expands the perceptual field of the network without changing the number of channels and extracts deeper features of the feature map. The output feature map of the point-by-point convolution is added point-by-point to the input feature map to obtain the output feature map. As shown in Figure 2B, depthwise convolution is formed based on depth-separable convolution, which has a lower number of parameters and lower cost compared to depth-separable convolution. As shown in Figure 2, assuming that U and O are the numbers of input and output channels, respectively, and k2is the size of the convolution kernel, the number of parameters of depthwise separable convolution is

FIGURE 2
www.frontiersin.org

Figure 2 Depth separable convolution and Depthwise convolution. (A) Depthwise separable convolution; (B) Depthwise convolution.

k2×U+U×O , and the number of parameters of depthwise convolution is

k2×U. Therefore, the ratio of parameters can be calculated as follows:

k2×U+U×Ok2×U=1+Ok2(5)

In this study, the DCWP module was designed with depthwise convolution as the backbone. To avoid missing feature information in the feature map compression process, no nonlinear activation function was used, and the Phish function was used to splice after the depthwise convolution to make full use of the relevant information in the feature map to transfer the shallow feature information to the deep layer. The DCWP module has been designed to be used as a tool for deeper layers; its structure is shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3 Depth Convolution with Phish Module.

3.5 Light residual module

To address the challenge of difficult deployment of CNN models for tomato leaf disease identification in mobile terminals and low-memory devices, there is an urgent need to design a new network architecture that guarantees high accuracy in identifying tomato leaf diseases in complex contexts while further reducing the total number of parameters and model size. We designed a residual-based LR module to make the model more lightweight and improve its deployment capability.

First, a convolutional layer and the Phish activation function were connected, and a batch normalization layer was added after the Phish activation function, as the network depth poses a challenge to the convergence speed of the entire network. Then, the sequentially connected convolutional layers, Phish activation function, and batch normalization layer were multiplexed in series to effectively improve the feature extraction for tomato leaf disease. Finally, residual connections constructed efficient residual blocks, which improved the ability of gradient propagation across layers and could prevent gradients in deep convolutional layers from vanishing. The structure of the LR module is shown in Figure 4.

FIGURE 4
www.frontiersin.org

Figure 4 Light Residual Module.

4 Experimental results and discussion

4.1 Hyperparameters and experimental equipment

The hyperparameter settings used in the experiments are presented in Table 3. The programming language used for the experiments was Python, and PyTorch was used as the deep learning framework. A detailed configuration of the development environment is presented in Table 4.

TABLE 3
www.frontiersin.org

Table 3 Hyperparameter configuration of CNN.

TABLE 4
www.frontiersin.org

Table 4 Software and hardware configuration.

To evaluate the performance of the model, recall, accuracy, precision and F1-score were selected as the metrics for comprehensive evaluation. The mathematical expressions for the evaluation metrics are as follows:

Recall=TPTP+FN(5)
Accuracy=TP+TNTP+TN+FP+FN(6)
Precision=TPTP+FP(7)
F1score=2TP2TP+FP+FN(8)

where TN is the number of true-negative samples, TP is the number of true-positive samples, FN is the number of false-negative samples, and FP is the number of false-positive samples.

4.2 Ablation experiment

Ablation experiments were used to demonstrate the effects of the DCWP and LR modules, and the results are presented in Table 5. The model identification accuracy without the DCWP and LR modules was 68.6%. The introduction of the DCWP module increased the accuracy of the model to 79%, and the introduction of the LR module increased the model accuracy to 97.5%; the model accuracy and F1-score were also significantly improved. The recognition accuracy of the proposed model with the DCWP and LR modules reached 99.3%, which was much higher than the recognition accuracy of the benchmark model. After both the DCWP and LR modules was added to the proposed model, the number of parameters of the model increased to 1.577 M, ensuring very few required parameters while maintaining the performance of image feature extraction from images of diseased tomato leaves.

TABLE 5
www.frontiersin.org

Table 5 Results of the ablation experiment.

4.3 Comparison results of the proposed model with classical CNN models

The proposed LightMixer model was then compared with other classical CNN models, namely AlexNet (Krizhevsky, 2014), ResNext (Xie et al., 2017), ResNet34 (He et al., 2016), ResNet50 (He et al., 2016), VGG16 (Simonyan and Zisserman, 2014), and VGG19 (Simonyan and Zisserman, 2014), in a comparative experiment on the tomato leaf disease dataset. Table 6 presents the comparative results for the different models. The proposed LightMixer model achieved 99.3% accuracy compared with other classical network models, which exhibited optimal performance. The weighty models of the suboptimal models are ResNext, ResNet34, and ResNet50. The accuracy of the ResNext, ResNet34, and ResNet50 models was stable at 99%. The accuracies of the AlexNet, VGG16, and VGG19 models were 96.2%, 96.6%, and 95.7%, respectively; they were approximately 3% less accurate and exhibited poorer performance than the LightMixer model. In addition, the LightMixer model proposed in this study has the fewest parameters among the many classical CNN models compared. The experimental results showed that the proposed LightMixer model has better performance on the tomato leaf disease dataset than other classical CNN models.

TABLE 6
www.frontiersin.org

Table 6 Comparison of the classical CNN models with the proposed LightMixer model.

Figure 5 shows the performances of the LightMixer model and other classical CNN models during training, including the accuracy and loss between these models. As shown in Figure 5A, the accuracy of the LightMixer model was higher than that of the other models at the initial stage. In addition, the LightMixer model was the first to transition to the smoothing curve and achieved 99% accuracy after the curve is smoothed. In addition to accuracy, the LightMixer model demonstrated strong convergence and generalization during the training process. As shown in Figure 5B, the loss curve of the LightMixer model was the first to transition to a smoothed curve and then remained stable. However, this was not the case for most other models, especially the AlexNet model. These comparisons demonstrate the superiority of the LightMixer model.

FIGURE 5
www.frontiersin.org

Figure 5 Training curves of the classical CNN models and the proposed LightMixer model. (A) Accuracy; (B) loss.

To observe the performance of different models in identifying different classes of tomato leaf diseases, a confusion matrix was used to visualize the classification results on the test set. Figure 6 shows the confusion matrix of the LightMixer model with other classical CNN models. The vertical axis represents the True Label, and the horizontal axis represents the Predicted Label. The values on the diagonal line represent the ratio of the number of correctly predicted disease samples for that class to the total number of samples for that class of images, and the values on the non-diagonal line represent the fraction of incorrect predictions. The values on the diagonal of the confusion matrix in Figure 6 Figure 5G are all very high, with more than half of the tomato leaf disease classes identified with an accuracy of 99% or higher. The LightMixer model was more than 95% accurate for the other disease categories. The confusion matrices for the ResNext and ResNet50 models were similar. However, the VGG19 model was the least effective in classifying all types of tomato leaf diseases, especially in identifying early blight, with an accuracy of only 77%. The experimental results showed that the LightMixer model can efficiently and accurately classify various types of tomato leaf diseases.

FIGURE 6
www.frontiersin.org

Figure 6 Confusion matrix between the classical CNN models and the proposed LightMixer model. (A) ALexNet; (B) ResNext; (C) ResNet34; (D) ResNet50; (E) VGG16; (F) VGG19; (G) LightMixer.

4.4 Comparison of the proposed model with the classical lightweight CNN model

Table 7 shows the identification results of the proposed model compared with those of the classical lightweight CNN model for the tomato leaf disease validation set. LightMixer demonstrated greater accuracy than the compared lightweight CNNs, including DenseNet121 (Huang et al., 2017), EfficientNet (Tan and Le, 2019), MobileNetV2 (Sandler et al., 2018), MobileNetV3 (Howard et al., 2019), ShuffleNetV2 (Ma et al., 2018), SqueezeNet (Iandola et al., 2016) and MNASNet (Tan et al., 2019)Although the Dense-Net121 and LightMixer models have similar accuracies, the LightMixer model requires approximately six times fewer parameters than the Dense-Net121 model. In addition, the proposed LightMixer model has only 1.5 M parameters, which is several to ten times less than the parameters of other lightweight models such as MobileNet V1, V2, and V3. Although the number of parameters of the SqueezeNet model was similar to that of the LightMixer model, the LightMixer model was more accurate. These experimental results demonstrate the advanced performance of the LightMixer model as a lightweight model, which is more suitable for lightweight execution of tomato leaf disease identification tasks.

TABLE 7
www.frontiersin.org

Table 7 Comparison of the classical lightweight CNN models with the proposed LightMixer model.

Figure 7 shows the training curves of the LightMixer model compared with those of other classical lightweight CNN models. Compared with other classical lightweight convolutional networks, the LightMixer model exhibits excellent learning ability during training. As shown in Figure 7A, the LightMixer model is one of the first models to transition to a smooth curve. In addition, it has the best accuracy and loss, as well as the least curve fluctuation among the different lightweight CNNs. Figure 7B shows the loss curves of different lightweight convolutional networks.

FIGURE 7
www.frontiersin.org

Figure 7 Training curves of the classical lightweight CNN models and the proposed LightMixer model. (A) Accuracy; (B) loss.

Figure 8 Figure 8 shows the confusion matrix of several lightweight CNNs on the test dataset. For the test set of tomato leaf diseases, the MNASNet model showed the worst classification results, where the accuracy of mosaic virus recognition was only 86%. In addition, the SqueezeNet model, which has a similar number of parameters to the LightMixer model, did not show an advantage in terms of classification effectiveness. The SqueezeNet model showed uneven classification accuracy for various types of tomato leaf diseases, with the lowest recognition accuracy of 93% for early blight, whereas the LightMixer model was 95% accurate for this type of disease. These experimental results demonstrate that the LightMixer model maintains its classification advantage for different types of tomato leaf diseases compared with other lightweight models.

FIGURE 8
www.frontiersin.org

Figure 8 Confusion matrix between the classical lightweight CNN models and the proposed LightMixer model. (A) DenseNet121;(B) EfficientNet;(C) MobileNetV2;(D) MobileNetV3;(E) ShuffleNetV2;(F) SqueezeNet;(G) MNASNet; (H) LightMixer.

5 Conclusions

The identification and classification of various tomato leaf diseases by CNN models are highly accurate; however, mainstream CNN models require higher computational and storage costs. In this study, an extremely lightweight CNN model called LightMixer is proposed for the automatic identification of tomato leaf diseases. In the proposed model, LR residual blocks are introduced as one of the main core units for building the backbone network. The architecture of LR residual blocks is designed to achieve better performance while requiring less computation and storage costs. The proposed LightMixer model also has a DCWP module that enhances the feature extraction capability of the network, highlights the nonlinear representation of the model, and compresses the parametric number of the model to enhance its classification capability. The combination of the LR residual block and DCWP module facilitates the ability of the model to utilize the feature map to better capture the disease features and help optimize the performance of the model. Experimental results demonstrate that the proposed model has 99.3% recognition accuracy in diagnosing tomato leaf diseases and has only 1.5 M parameters, which is better than other tested classical models. In the next step, the deployment of the proposed model in mobile or embedded device environments will be our main goal to help farmers accurately identify and detect tomato leaf diseases. Further exploration of the generalizability of the proposed model to detect and identify a variety of other plant diseases will be part of our future plans.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/datasets/abdallahalidev/plantvillage-dataset.

Author contributions

MT contributed to conception and design of the study. YZ designed and performed the experiment, managed the algorithms and result analysis and wrote the manuscript. ZT participated in the experiments and manuscript revision. All authors contributed to the article and approved the submitted version.

Acknowledgments

We are appreciative of the reviewers’ valuable suggestions on this manuscript and the editor’s efforts in processing the manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2023.1166296/full#supplementary-material

Supplementary Figure 1 | Samples of tomato diseased leaves. (A) Bacterial spot; (B) Early blight; (C) Healthy; (D) Late blight; (E) Leaf mold; (F) Septoria leaf spot; (G) Two-spotted spider mite; (H) Target spot; (I) Mosaic virus; (J) Yellow Leaf Curl Virus.

Supplementary Figure 2 | The result after data enhancement. (A) Original image; (B) Random Image Crop; (C) horizontal random flip; (D) vertical random flip; (E) random rotation.

References

Abbas, A., Jain, S., Gour, M., Vankudothu, S. (2021). Tomato plant disease detection using transfer learning with c-GAN synthetic images. Comput. Electron. Agric. 187, 106279. doi: 10.1016/j.compag.2021.106279

CrossRef Full Text | Google Scholar

Agarwal, M., Gupta, S. K., Biswas, K. (2020). Development of efficient CNN model for tomato crop disease identification. Sustain. Computing: Inf. Syst. 28, 100407. doi: 10.1016/j.suscom.2020.100407

CrossRef Full Text | Google Scholar

Al‐Gaashani, M. S., Shang, F., Muthanna, M. S., Khayyat, M., Abd El-Latif, A. A. (2022). Tomato leaf disease classification by exploiting transfer learning and feature concatenation. IET Image Process. 16, 913–925. doi: 10.1049/ipr2.12397

CrossRef Full Text | Google Scholar

Arun Pandian, J., Gopal, G. (2019). Data from: identification of plant leaf diseases using a 9-layer deep convolutional neural network (Mendeley Data). Available at: https://data.mendeley.com/datasets/tywbtsjrjv/1.

Google Scholar

Barman, U., Choudhury, R. D., Sahu, D., Barman, G. G. (2020). Comparison of convolution neural networks for smartphone image based real time classification of citrus leaf disease. Comput. Electron. Agric. 177, 105661. doi: 10.1016/j.compag.2020.105661

CrossRef Full Text | Google Scholar

Bhagat, M., Kumar, D., Haque, I., Munda, H. S., Bhagat, R. (2020). “Plant leaf disease classification using grid search based SVM,” in 2nd International Conference on Data, Engineering and Applications (IDEA). (Bhopal, India: IEEE), 1–6. doi: 10.1109/IDEA49133.2020.9170725

CrossRef Full Text | Google Scholar

Bhatia, A., Chug, A., Singh, A. P. (2020). “Hybrid SVM-LR classifier for powdery mildew disease prediction in tomato plant,” in 2020 7th International conference on signal processing and integrated networks (SPIN). (Noida, India: IEEE), 218–223. doi: 10.1109/SPIN48934.2020.9071202

CrossRef Full Text | Google Scholar

Bi, C., Wang, J., Duan, Y., Fu, B., Kang, J.-R., Shi, Y. (2022). MobileNet based apple leaf diseases identification. Mobile Networks Appl. 25, 1–9. doi: 10.1007/s11036-020-01640-1

CrossRef Full Text | Google Scholar

Cap, Q. H., Tani, H., Kagiwada, S., Uga, H., Iyatomi, H. (2021). LASSR: effective super-resolution method for plant disease diagnosis. Comput. Electron. Agric. 187, 106271. doi: 10.1016/j.compag.2021.106271

CrossRef Full Text | Google Scholar

Chen, W., Chen, J., Duan, R., Fang, Y., Ruan, Q., Zhang, D. (2022). MS-DNet: a mobile neural network for plant disease identification. Comput. Electron. Agric. 199, 107175. doi: 10.1016/j.compag.2022.107175

CrossRef Full Text | Google Scholar

Chen, J., Chen, J., Zhang, D., Sun, Y., Nanehkaran, Y. A. (2020). Using deep transfer learning for image-based plant disease identification. Comput. Electron. Agric. 173, 105393. doi: 10.1016/j.compag.2020.105393

CrossRef Full Text | Google Scholar

Chen, J., Zhang, D., Zeb, A., Nanehkaran, Y. A. (2021). Identification of rice plant diseases using lightweight attention networks. Expert Syst. Appl. 169, 114514. doi: 10.1016/j.eswa.2020.114514

CrossRef Full Text | Google Scholar

Chopda, J., Raveshiya, H., Nakum, S., Nakrani, V. (2018). “Cotton crop disease detection using decision tree classifier,” in 2018 International Conference on Smart City and Emerging Technology (ICSCET). (Mumbai, India: IEEE). pp. 1–5. doi: 10.1109/ICSCET.2018.8537336

CrossRef Full Text | Google Scholar

Darwish, A., Ezzat, D., Hassanien, A. E. (2020). An optimized model based on convolutional neural networks and orthogonal learning particle swarm optimization algorithm for plant diseases diagnosis. Swarm evolutionary Comput. 52, 100616. doi: 10.1016/j.swevo.2019.100616

CrossRef Full Text | Google Scholar

Fan, X., Luo, P., Mu, Y., Zhou, R., Tjahjadi, T., Ren, Y. (2022). Leaf image based plant disease identification using transfer learning and feature fusion. Comput. Electron. Agric. 196, 106892. doi: 10.1016/j.compag.2022.106892

CrossRef Full Text | Google Scholar

Faostat, D. (2022). “Crops and livestock products,” in Statistics division, food and agriculture organization of the united nations(Rome, Italy: Food and Agriculture Organization of the United Nations).

Google Scholar

He, K., Zhang, X., Ren, S., Sun, J. (2016). Deep residual learning for image recognition. Proc. IEEE Conf. Comput. Vision Pattern recognition, 770–778. doi: 10.1109/CVPR.2016.90

CrossRef Full Text | Google Scholar

Hendrycks, D., Gimpel, K. (2016). Gaussian Error linear units (gelus). arXiv preprint arXiv 1606, 08415. doi: 10.48550/arXiv.1606.08415

CrossRef Full Text | Google Scholar

Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., et al. (2019). “Searching for mobilenetv3,” in Proceedings of the IEEE/CVF international conference on computer vision. (Seoul, Korea (South): IEEE),1314–1324.

Google Scholar

Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K. Q. (2017). “Densely connected convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Honolulu, Hawaii: IEEE), 4700–4708.

Google Scholar

Iandola, F. N., Han, S., Moskewicz, M. W., Ashraf, K., Dally, W. J., Keutzer, K. (2016). SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and< 0.5 MB model size. arXiv preprint arXiv 1602, 07360. doi: 10.48550/arXiv.1602.07360

CrossRef Full Text | Google Scholar

Kamal, K., Yin, Z., Wu, M., Wu, Z. (2019). Depthwise separable convolution architectures for plant disease classification. Comput. Electron. Agric. 165, 104948. doi: 10.1016/j.compag.2019.104948

CrossRef Full Text | Google Scholar

Karthik, R., Hariharan, M., Anand, S., Mathikshara, P., Johnson, A., Menaka, R. (2020). Attention embedded residual CNN for disease detection in tomato leaves. Appl. Soft Computing 86, 105933. doi: 10.1016/j.asoc.2019.105933

CrossRef Full Text | Google Scholar

Krizhevsky, A. (2014). One weird trick for parallelizing convolutional neural networks. arXiv preprint arXiv 1404, 5997. doi: 10.48550/arXiv.1404.5997

CrossRef Full Text | Google Scholar

Lakshmanarao, A., Babu, M. R., Kiran, T. S. R. (2021). “Plant disease prediction and classification using deep learning ConvNets,” in 2021 International Conference on Artificial Intelligence and Machine Vision (AIMV). (Gandhinagar, India: IEEE). pp. 1–6. doi: 10.1109/AIMV53313.2021.9670918

CrossRef Full Text | Google Scholar

Lee, S. H., Goëau, H., Bonnet, P., Joly, A. (2020). Attention-based recurrent neural network for plant disease classification. Front. Plant Sci. 11, 601250. doi: 10.3389/fpls.2020.601250

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, Y., Hu, Y., Cai, W., Zhou, G., Zhan, J., Li, L. (2022). DCCAM-MRNet: mixed residual connection network with dilated convolution and coordinate attention mechanism for tomato disease identification. Comput. Intell. Neurosci. 2022, 15. doi: 10.1155/2022/4848425

CrossRef Full Text | Google Scholar

Ma, N., Zhang, X., Zheng, H.-T., Sun, J. (2018). “Shufflenet v2: practical guidelines for efficient cnn architecture design,” in Proceedings of the European conference on computer vision (ECCV). (Munich, Germany: Springer), 116–131.

Google Scholar

Nations, F. A. A. O. O. T. U. (2020) Crops and livestock products. Available at: https://www.fao.org/faostat/en/#data/QCL/visualize (Accessed 3 March 2023).

Google Scholar

Naveen, P. (2022). Phish: a novel hyper-optimizable activation function.

Google Scholar

Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C. (2018). “Mobilenetv2: inverted residuals and linear bottlenecks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Salt Lake City, UT, USA: IEEE), 4510–4520.

Google Scholar

Simonyan, K., Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv 1409, 1556. doi: 10.48550/arXiv.1409.1556

CrossRef Full Text | Google Scholar

Singh, P., Verma, A., Alex, J. S. R. (2021). Disease and pest infection detection in coconut tree through deep learning techniques. Comput. Electron. Agric. 182, 105986. doi: 10.1016/j.compag.2021.105986

CrossRef Full Text | Google Scholar

Tan, M., Chen, B., Pang, R., Vasudevan, V., Sandler, M., Howard, A., et al. (2019). “Mnasnet: platform-aware neural architecture search for mobile,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. (Long Beach, CA, USA: IEEE), 2820–2828.

Google Scholar

Tan, M., Le, Q. (2019). “Efficientnet: rethinking model scaling for convolutional neural networks,” in International conference on machine learning. 6105–6114.

Google Scholar

Xie, S., Girshick, R., Dollár, P., Tu, Z., He, K. (2017). “Aggregated residual transformations for deep neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Honolulu, HI, USA: IEEE),1492–1500.

Google Scholar

Xu, C., Ding, J., Qiao, Y., Zhang, L. (2022). Tomato disease and pest diagnosis method based on the stacking of prescription data. Comput. Electron. Agric. 197, 106997. doi: 10.1016/j.compag.2022.106997

CrossRef Full Text | Google Scholar

Zeng, W., Li, H., Hu, G., Liang, D. (2022). Lightweight dense-scale network (LDSNet) for corn leaf disease identification. Comput. Electron. Agric. 197, 106943. doi: 10.1016/j.compag.2022.106943

CrossRef Full Text | Google Scholar

Zhao, S., Peng, Y., Liu, J., Wu, S. (2021). Tomato leaf disease diagnosis based on improved convolution neural network by attention module. Agriculture 11, 651. doi: 10.3390/agriculture11070651

CrossRef Full Text | Google Scholar

Keywords: tomato leaf disease, lightweight model, convolutional neural networks, deep learning, disease detection

Citation: Zhong Y, Teng Z and Tong M (2023) LightMixer: A novel lightweight convolutional neural network for tomato disease detection. Front. Plant Sci. 14:1166296. doi: 10.3389/fpls.2023.1166296

Received: 15 February 2023; Accepted: 06 April 2023;
Published: 09 May 2023.

Edited by:

Mansour Ghorbanpour, Arak University, Iraq

Reviewed by:

Nisha Pillai, Mississippi State University, United States
Yourong Chen, Zhejiang Shuren University, China

Copyright © 2023 Zhong, Teng and Tong. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mengjun Tong, 20130035@zafu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.