- 1School of Mechanical Engineering, Tongling University, Tongling, China
- 2School of Computer and Software Engineering, Anhui Institute of Information Technology, Wuhu, China
- 3College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou, China
Tea winnowing is a key process in tea processing. At present, tea winnowing parameters are adjusted by manual observation of tea leaves. This results in the uncertainty of winnowing quality. In this work, we propose a new tea winnowing method based on deep learning for the characteristics of white tea. Firstly, the YOLOv11 model is improved by introducing ACmix and EUCB. The recognition accuracy of the improved YOLO-AE model is improved by 2.1%, and the detection time is shortened by 40%, which significantly improves the detection performance and shortens the inference time. The region segmentation and convolution neural network algorithm are used to distinguish the proportion parameters of each grade in tea in real time, and the accurate wind selection parameters are obtained by combining the winnowing theory. The recognition accuracy of the verification set of the recognition model attains 94%. The MAP (0.5:0.95) is 0.93. A test on the tea winnowing parameter control test bench reveals that the identification accuracy of tea materials with different proportions is consistent. Additionally, the difference between the two batches of high-quality white tea is less than 3%. The winnowing scheme proposed in this study can provide the basic theory and technical support for the design of tea precision winnowing equipment.
1 Introduction
Tea is one of the important economic crops in China. With the improvement in living standards, the demand for high-quality tea has increased (Liu, 2019). Tea production is considerable in summer and autumn. However, because tea is mixed with more impurities such as tea stems and broken leaves, the price is low. The impurities in summer and autumn tea can be removed by pneumatic separation. The price of tea after removing impurities is 3–5 times that of the raw materials (Gan et al., 2018). However, the pneumatic separation parameters required for different types of tea or different batches of the same type of tea vary. If the pneumatic separation parameters are not adjusted accurately in a timely manner, the quality of pneumatic separation would decrease rapidly.
At present, the tea sorted by a color sorter displays high precision. However, the efficiency of sorting cannot satisfy the requirements of mass production of tea, and the price is high. The sorting of large quantities of tea relies mainly on pneumatic separation equipment. However, the current structural parameters of tea pneumatic separation rely on manual experience design. Workers cannot correct the air separation parameters in a timely manner according to the different raw materials of tea. This results in inconsistent quality of different batches of tea.
A substantial amount of research has been conducted on tea classification worldwide. The tea sample information was obtained by spectroscopy. Then, the SVM model was optimized by particle swarm optimization, artificial bee colony, genetic algorithm, and other algorithms. Finally, an accurate classification of tea was realized (Mishra et al., 2019; Hussain et al., 2021; Mingyin et al., 2021; Yuhan et al., 2022). At the same time, some scholars use YOLO model to identify tea buds, which has a good recognition effect in complex background (Lyu et al., 2021; Xu et al., 2022; Cheng et al., 2023; Dong et al., 2024). The physical parameters of tea, such as area, perimeter, and diameter, can be used as indexes to distinguish the grade of tea by a machine vision system (Rahman et al., 2021). Jiang et al. used the symmetry detection algorithm to detect the edge of tea images. This further improved the efficiency and accuracy of tea classification (Jiang and Chen, 2020). Song et al. obtained the histogram of tea shape features. Through various machine learning model recognition techniques, it was determined that the model based on shape feature histogram is the best (Yan et al., 2018). Bakhshipour et al. extracted the color and texture features of four types of black tea. They observed that the neural network constructed by correlation feature selection had the best classification effect (Bakshipour et al., 2018). Sun et al. collected five types of hyperspectral images of green tea and established a support vector machine model for green tea recognition. The classification accuracy rate attained 96% (Sun et al., 2018). The above research results reveal that the deep learning algorithm has a good effect on tea recognition and classification. However, the experiment determined that the existing model has a low accuracy in identifying samples with similar tea. Therefore, it is urgent to develop an accurate identification and classification method for similar variations in tea parameters to prevent the adverse impact on the windward side and weight assessment in the subsequent classification and statistics.Zhang et al. studied the flow field (different wind speed distributions) of tea particles with different weights at different air duct outlets. The influence of drift trajectory, horizontal drift law, and wind direction angle on the quality of tea air separation (Zhang et al., 2013). Zhong et al. used the method of simulation analysis to study the trajectory of tea particles under different wind speed distributions in a tea air separator. When the wind speed is distributed according to upper small and lower large, the wind selection effect of tea particles of different quality is better. The impurity rate of each tea outlet can be reduced by increasing the spoiler (Zhing et al., 2013). Wu et al. designed a double-layer tea air separator and established an automatic control system based on a programmable logic controller. The separation baffle of a tea air separator was adjusted electrically using a frequency converter. Then, the optimum process parameters were obtained by an orthogonal test (Zhengmin et al., 2022). Luo et al. defined the simulation environment of tea material, flow field, and wind field wall by fluid-solid coupling. Through the simulation test, the wind selection parameters of tea were determined accurately, and the accuracy of tea wind selection was improved (Luo et al., 2022). The above researchers mainly simulated and analyzed the air separator through theoretical analysis and fluid dynamics methods to improve the structure of the pneumatic separation. The aim was to improve the air separation effect. Few studies have been conducted on how to configure the parameters of the air separator according to different tea types and impurity rates to realize intelligent air separation. Considering the practical application, the following problems persist: (1) The current research is focused on the slower environment, which cannot satisfy the real-time online recognition of the air separator. (2) The proportion of each grade of tea cannot be analyzed statistically in real time. (3) At present, there is no application of deep learning recognition combined with wind selection.
To address the demand for high efficiency and high precision for large quantities of tea grading, this study combined the previous research foundation of the fluid–solid coupling tea wind separation theory. A sorting scheme combining machine vision and air separator was proposed. First, the identification model of different components of tea was established, and the grade proportion of export tea was analyzed online in real time. Combined with the adjustment model of winnowing parameters, the accurate winnowing parameters required for current tea were determined. Then, the air separation parameters of the air separator were adjusted in real time so that different tea materials can maintain the best separation parameters in real time. The specific summary is as follows. Because the tea in this study is a small target, and the characteristics of different types of tea are similar, the recognition accuracy of the model is improved by further optimizing YOLOV11, and finally the accurate and efficient winnowing is realized.
2 Materials and methods
2.1 Dry tea material parameters
The experimental material used in this study was white tea raw material. The preliminary processing technology of white tea involves picking, withering (withering + screening), drying, and other processes. After the preliminary processing of tea. White tea products are formed through fine processing (picking, stacking, blending, stacking, redrying, and packing). After the initial processing of the white tea, the main impurities in the mixed material are yellow flakes and broken powder. Fine processing can remove most of the yellow flakes and debris impurities by air separation. The special white tea is selected from the secondary white tea by sorting. The main components of the test materials selected in this study were high-quality white tea (all single buds, used to make white silver needles), secondary white tea (mainly a bud and a leaf, a bud and two leaves, used to make white peony), and yellow flake impurities (Figure 1).
2.2 Winnowing platform based on deep learning
For the fine processing of white tea. The problem of poor air separation effect caused by inaccurate manual setting of air separator parameters. In this study, a real-time control experimental platform for tea selection parameters based on deep learning was designed. It is illustrated in Figure 2. The proportion of tea components at the outlet position is calculated in real time by the recognition model. Then, the actuator is controlled to adjust the air separation parameters in real time. Its overall structure is shown in Figure 2A. The winnowing platform is mainly composed of a tea recognition system, separator position control system, and winnowing mechanism. The tea is entered uniformly into the internal of the air separator through the vibrating feeder. The fan produces air flow into the fan horizontally. The tea drifts to the right under the action of the air flow inside the wind separator. Different types of materials in the tea material fall in different grilles below. Tea slides through the identification area through the slideway. The camera directly above the slideway captures the image. The tea identification system to determine the impurities of tea and related parameters to identify the current batch of tea require winnowing parameters. The specific structure is shown in Figure 2B. The control system adjusts the wind speed, grille position, and feeding amount of the fan to achieve the best separation quality.
Figure 2. Winnowing parameters real-time adjustment test bench [(A) whole machine diagram; (B) recognition system structure diagram].
2.3 Real-time recognition model of white tea material proportion
2.3.1 White tea sample data collection
For more accurate results of white tea material category identification and analysis, the captured tea image needs to be clear and the background contrast needs to be high. Images with clear contours are convenient for manual labeling and for machine learning and recognition. The test bench selects the strip LED light source as the light source of this test. The image acquisition device is an industrial camera (acA1920-150uc, Germany). The size of the shooting area is 200 × 100 mm. The shooting distance is 220 mm. The discharge port adopts a blue background plate. Blue is more conducive to the shooting of tea images, reducing the exposure of images, preventing shadows in images, and improving the quality of images than other color background plates. Through the image acquisition platform, a total of 1095 sample images were collected in this experiment. These include 426 high-quality white tea images, 462 secondary white tea samples, and 207 yellow chips (yellow chips, broken leaves, and tea chips) samples. To improve the diversity of datasets to adapt to more complex environments, this study increased the number of images in the dataset by color adjustment and mirror flipping of images. After amplifying 1095 datasets, 5756 images were obtained as the final dataset. Among the 5756 samples, the training, validation, and test sets accounted for 70%, 20%, and 10%, respectively.
2.3.2 White tea category recognition model
Faster R-CNN, SSD, and YOLO are the most widely used object detection models. The defects of YOLOV11 model in identifying small tea targets are improved.In this study, four models were used to train the white tea grade recognition model. First, the dataset constructed in-house in this study was used to input the three models to generate the corresponding level recognition model. Then, the results were compared and analyzed. Finally, the optimal model was selected to construct the white tea grade recognition model in this study.
The Faster R-CNN model is based on the R-CNN and Fast R-CNN network structure. The model mainly combines feature extraction, candidate box extraction, bounding box regression, and classification. This considerably improves the detection speed. The network structure is shown in Figure 3.
The backbone network used by the SSD model is VGG-16, as shown in Figure 4. It is the basic framework of the VGG-16 model. The feature extraction network of the SSD network model is optimized for VGG-16.
The tea that needs to be identified in this study is a small target, and it needs a higher speed when it is identified, and it also needs to be lightweight when deployed on the device. Compared with YOLOV5 and YOLOV8, YOLOV11 introduces C3K2/REP module to reduce inference delay. Integrated attention mechanism to improve global perception. At the same time, YOLOV11 is more suitable for the pursuit of accuracy and real-time scenes. Therefore, this study introduces a C2PSA-ACmix hybrid module at the end of the backbone network of YOLOV11, which combines the advantages of self-attention mechanism and convolution operation. ACmix first uses 1x1 convolution to project the input feature map to generate a set of intermediate features, and then reuses and aggregates these intermediate features according to different paradigms. Aiming at the difficulty of detecting dense small targets such as tea, a feature enhancement module (EUCB) is introduced at the key position of the head network. The workflow includes: 1.Bilinear upsampling (2 ×) expands the spatial resolution of the feature map. The DWC-BN-ReLU sequence achieves efficient feature nonlinear transformation. 3.1×1convolution completes channel dimension simplification to ensure compatibility with downstream networks. The self-attention module gives the model long-range dependent modeling ability, while the depthwise separable convolution (DWC) effectively captures local texture details. The combination of the two significantly improves the robustness of feature representation. The network structure is shown in Figure 5.
The analysis of the ACmix module has two phases. In the first stage, three 1 × 1 convolution projections are performed on the input feature map. In the second stage, the outputs of the two paths of Convolution aggregation and self-attention aggregation are added, and the intensity is controlled by two learnable scalars (Pan et al., 2022). The expression is shown in Equation 1.
The ECUB module is based on the attention mechanism. The core idea is to extract features by deep separable convolution DWC (.) Then the batch normalization operation BN (.) is performed to stabilize the training process. The nonlinear transformation is introduced by the ReLU (.) activation function. Finally, 1 × 1 convolution is used to compress the number of channels in the feature map. Make it match with the input requirements of the next stage (Rahman et al., 2024). The expression is shown in the Equation 2.
2.3.3 Identify model evaluation indicators
The computer used for model training is configured as a CPU processor (12th Gen Intel (R) Core (TM) i7-12700H 2.70 GHz). The graphics card is NVIDIA RTX3060 GPU, and the memory is of 16 GB. The operating system running the model is Windows 11 Professional 64 bit. The software tool is PyCharm 2021.3.2. The experiment was implemented in the Pytorch framework. To evaluate the effect of the model in identifying the tea shoots, this study used the recognition accuracy, recall rate, and map as the evaluation indexes to evaluate the training recognition model. The specific calculation indexes are shown in Equations 3–5.
Among them, TP is the number of correctly identified white tea categories, FP is the number of misidentified white tea categories, and FN is the number of unidentified white tea categories. For the tea target detection task in this study, the definition of TP and FP is based on the following two criteria: (1). The category of the prediction box must be consistent with the category of the real target. (2). IOU threshold: The IOU value between the prediction box and the real target box is greater than 0.5. If the prediction box satisfies the above two conditions, it is regarded as TP. If the prediction box does not meet any of the above conditions, it is considered FP.The larger the values of PC, RC and FC, the higher the accuracy of the model to identify the white tea category. The time required for the tea to enter the camera field-of-view to obtain the type information of the white tea category by the recognition model also reflects the superiority of the model from the side. The shorter the average processing time, the better is the model.
2.4 Statistical analysis algorithm of white tea material category
When the white tea recognition model detects different types of tea materials, the model also needs to count the number and area proportion of different tea material types. The unit number and area of tea determine the setting of wind selection parameters. Therefore, in this study, the number and area of different materials were determined by region segmentation, and the volume of different materials was estimated by combining the shape characteristics of different materials. The weight was determined based on a statistical analysis of the density of different types of materials. The material statistical analysis process is shown in Figure 6.
The first step was to use the camera to obtain the tea photographs at the outlet of the tea air separator. Tea recognition tests were performed using the trained faster R-CNN, SSD, YOLO, and YOLO-AE models. Finally, the best recognition model was selected for the tea recognition experiment. The second step was to use the marker box to indicate the tea types identified by each model. Then, the region segmentation algorithm was used to extract the tea image in the marker box region (Wang et al., 2024; Chen et al., 2025). The third step was to extract the images of the R, G, B, H, S, and V channels from the segmented tea images. The channel with the highest contrast was selected for the area ratio calculation. Then, the image threshold segmentation of the tea S channel was extracted to the tea area. The areas of the tea area and marker box were obtained by area calculation. The ratio of the two was used as the ratio of the tea area.
Ten tea materials were selected randomly. The density and thickness parameters of different tea types were analyzed statistically. The specific results are shown in Table 1. The distance and area utilized by the camera are fixed. Therefore, the proportion of each tea can be determined by counting the number of various teas and then calculating the total proportion of the area occupied by the images. By counting the density and thickness of different types of tea, the weight can be calculated according to the measured tea area. By calculating the tea area and weight, the adjustment of the wind selection parameters was determined.
2.5 Winnowing parameter adjustment mechanism
After the identification system determines the proportion of different types of tea materials. According to the previous research results of wind field simulation, the size of the blanking port is adjusted by adjusting the left and right position parameters of the partition plate. The moving mechanism of the dividing plate is shown in Figure 7. The partition plate is fixed on the slider. The slider moves toward the left and right through the rotation of the screw. One end of the screw is connected with the stepping motor. The positions of the partition plate are left, right, and origin restriction. The reference point of the dividing plate is the origin when adjusting the position each time. The left and right limits ensure that the partition plate does not exceed the limit position.
The intelligent control system of tea air separator is shown in Figure 8. The host computer control system integrates related information such as the fan frequency, feeder frequency, baffle position, current wind speed, and tea image at the discharge port. The parameters in the process of air separation can be observed and recorded in real time through the upper computer control system. Simultaneously, the control system can be controlled manually and automatically.
3 Results and discussion
3.1 Model training results
The white tea grade data set was imported into the Faster R-CNN model for training. For the small target detection of white tea in this study, the batch size was set to nine. The results obtained after training are shown in Figure 9. Figure 9A shows the average accuracy (MAP) of different types of tea when the tea category is 0,1,2. Figure 9B shows the log _ average miss rate when the tea category is 0,1,2. The smaller the value, the better is the model. Figure 9C displays the training set loss function curve and validation set loss function curve. Figures 9D–F represent the accuracy of different grades of white tea at IOU = 0.5. Figures 9G–I present the recall rate of different grades of white tea at IOU = 0.5.
Figure 9. Faster R-CNN model test results [(a) tea average accuracy; (b) tea log _ average miss rate; ((c–e): faccuracy of different grades of white tea at IOU = 0.5 ; (g–i): the recall rate of different grades of white tea at IOU = 0.5].
The white tea grade dataset was imported into the SSD model for training. The number of iterations in this experiment was 150. The batch size was set to eight for the small target detection of white tea in this study. The results obtained after training are shown in Figure 10. Figure 10A shows the average accuracy of the model to identify tea when the tea category is 0,1,2. Figure 10B shows the log _ average miss rate when the tea category is 0,1,2. The smaller the value, the better is the model training effect. Figure 10C displays the variation trend of the loss function curves of the training and validation sets with the increase in the number of training times. It can be observed from the figure that the model approaches a smooth state after 100 times of training. Figures 9D–F represent the accuracy of the model at IOU = 0.5 when the tea category is 0,1,2. Figures 10G–I present the recall rate of different grades of white tea at IOU = 0.5.
Figure 10. SSD model test results [(a) tea average accuracy; (b) tea log _ average miss rate; (c–e) faccuracy of different grades of white tea at IOU = 0.5 ; (g–i): the recall rate of different grades of white tea at IOU = 0.5].
The white tea grade dataset was imported into the YOLOV11 model for training. The number of iterations of this training was 300. Because this study was aimed at the tea target with a smaller target, the batch size was set to eight. After training, the experimental results of cls _ loss, recall, precision, MAP _ 0.5, MAP _ 0.5: 0.95 of the model are shown in Figure 11. After 75 iterations of model training, the curve is close to a stable trend. The cls _ loss of the model is close to 0.76. Precision approaches 0.77. The recall of the model approaches 0.86. The MAP _ 0.5 of the model is close to 0.84. The MAP _ 0.5: 0.95 of the model is close to 0.52.
Figure 11. Experimental results of YOLOv11 model [(a) model cls loss; (b) model recall; (c) model precision; (d) model mAP 0.5; (e) model mAP _ 0.5 : 0.95].
The white tea grade dataset was imported into the YOLO-AE model for training. The number of iterations of this training was 300. Because this study was aimed at the tea target with a smaller target, the batch size was set to eight. After training, the experimental results of cls _ loss, recall, precision, MAP _ 0.5, MAP _ 0.5: 0.95 of the model are shown in Figure 12. After 75 iterations of model training, the curve is close to a stable trend. The cls _ loss of the model is close to 0.47. Precision approaches 0.82. The recall of the model approaches 0.9. The MAP _ 0.5 of the model is close to 0.89. The mAP _ 0.5: 0.95 of the model is close to 0.52.
Figure 12. Experimental results for cls-loss in YOLO-AE model. [(a) model cls loss; (b) model recall; (c) model precision; (d) model mAP 0.5; (e) model mAP _ 0.5 : 0.95].
By training the four models, the corresponding model parameters were obtained. The test data of Precision, Recall, MAP _ 0.5, MAP _ 0.5: 0.95, and frame recognition time time for each model are shown in Table 2. The comparison results reveal that the weight file of the Faster R-CNN model is the largest. Moreover, these predict that an image requires the longest time. The SSD model displayed accuracy, recall, and MAP values lower than those for the other two models. Among the three models, the YOLO-AE model yielded the best training results. The model requires the shortest time to predict, the minimum weight file is 6.3 M, and the model is lightweight. In the identification, the shape parameters of different grades of tea are close, which requires high accuracy. The improved YOLOV11 model and the number of parameters and floating-point operations before the improvement only increased by 3%, but the recognition accuracy increased by 3%. Compared with the YOLOV5 model, the improved model parameters are reduced by 61%, and the recognition accuracy is increased by 3%. The accuracy of the system identification is the first pursuit factor. Based on the analysis of the above factors, this study chooses the improved YOLOV11 model to be more suitable for tea selection regulation.
Then, the experimental results were compared with those of Chuan et al (Yin et al., 2023). The recognition system in this study improved the overall recognition accuracy by an average of 3% through light adjustment, background plate selection, and dataset preprocessing.When tea leaves enter the recognition area, there is a small amount of overlap. In this study, tea diversion to avoid overlapping phenomenon. When the speed of tea entry is fast, the picture is blurred. The problem of image blurring is solved by increasing the frame rate of the camera. The tea part enters the camera field of view, resulting in misidentification. Based on the above factors, there are differences in the recognition accuracy of tea of the same grade. As shown in Figure 13 B1. The probability of this occurrence in the actual test is less than 3%, so the impact on the final test results can be ignored.
Figure 13. The verification result for tea grade recognition [(A1, B4, C2, C3, C4, D4) : category 0 recognition results ; (A2, A3, B2, B3, C1, D1, D2) : Category 1 recognition results ; (A4, B1, D3) : Category 2 recognition results].
3.2 Accuracy verification of YOLO-AE recognition system
The recognition accuracy of the recognition system based on the YOLO-AE model was verified. In this study, high-quality white tea in the white tea raw materials accounted for 85%, secondary white tea accounted for 10%, and broken leaf impurities accounted for 5% as test materials. The mixed tea was divided randomly into 10 parts. In turn, 10 tea leaves were put into the wind selection test bench for the identification test. The accuracy of the tea identification model was verified by 10 experiments. Finally, the feasibility of the system was analyzed based on the recognition time and accuracy. A set of recognition results is shown in Figure 13. The three tea categories were marked with three color border boxes. The number 0,1 and 2 on the left side of the label represent high-quality tea, secondary tea and. The values behind represent the recognition accuracy.There is a small amount of overlap in tea. In this study, tea diversion to avoid overlapping phenomenon. When the speed of tea entry is fast, the picture is blurred. The problem of image blurring is solved by increasing the frame rate of the camera. Incomplete tea leaves in the image lead to misidentification. Based on the above factors, there are differences in the recognition accuracy of tea of the same grade. As shown in Figure 13-B1. The probability of this occurrence in the actual test is less than 3%, so the impact on the final test results can be ignored.
The accuracy of the YOLO-AE model in identifying different types of tea was tested through 10 experiments. Finally, the statistical model recognition time, recognition accuracy of various grades of tea, and proportion of each component of tea were determined. The results are shown in Table 3. The results of the 10 experiments reveal that the recognition time for each image was less than 45 ms. The recognition accuracy for high-quality white tea, secondary white tea, and debris impurities was above 90%. The standard deviation of the recognition accuracy of the three kinds of tea is 9.73,9.72 and 9.75. The results show that the recognition accuracy is stable. The statistical analysis of the proportion of various teas was essentially consistent with the ratio set by oneself. The deviation of each tea proportion was ± 0.8%. Among these, the highest proportion of high-quality white tea raw materials had a better recognition effect (accounting for nearly 85%). Then, the experimental results were compared with those of Huang et al (Shaohua and Xifeng, 2022). The recognition accuracy of the recognition system in this study for various components of white tea materials was better than that of its improved model by 1–3%.
3.3 Tea winnowing system experiment
To verify the separation effect of the air separation system, this study compared two proportions of white tea materials. The proportions of the first high-quality white tea, secondary white tea, and debris impurities were 85%, 10%, and 5%, respectively. The second high-quality white tea, secondary white tea, and debris impurities accounted for 70%, 22%, and 8%, respectively. The first batch of raw materials was fed first, followed by the second batch. Then, the variations in fan frequency and separator position were compared with the first batch of tea. In this study, 10 repeated experiments were performed. After the tea was identified by the recognition system, the wind selection parameters were adjusted, and the proportion of high-quality white tea at the discharge outlet was counted manually. The results are shown in Table 4.
The winnowing parameters required for the first tea and second tea differed significantly. The feeding speed remained unaltered. The maximum wind speed difference of the fan was 2.7 m/s, and the minimum was 1.1 m/s. The difference in the moving position of the partition plate was positive, and the partition plate moved in the same direction. The maximum difference was 3.1 cm, and the minimum difference was 1.3 cm. The maximum difference in the content of high-quality white tea at the second outlet was 2.1%, and the minimum value was 0.6%. The difference in the content of high-quality white tea was less than 3%. This result shows that the recognition system can adjust the wind selection parameters in real time with the change of tea materials for winnowing of different batches of tea.
In the actual tea production, the dim light, high noise and hot and humid environment lead to the inability to stay for a long time. Therefore, the wind selection parameters cannot be adjusted in time and accurately. At the same time, manual adjustment takes a long time. The system can quickly adjust the parameters in time.The tea identification environment is a closed space. The light is stabilized by an industrial light source and will not be affected by other light. As the use time increases, the dust of the tea will cover the lens, resulting in unclear pictures. In the subsequent use process, it is necessary to regularly clean the dust on the lens to ensure the stability of the shooting. At present, we are establishing a database of impurity sorting of green tea. Through the training of this model and the simulation of air separation parameters, the impurity separation of other kinds of tea can be realized.
4 Conclusion
This study addressed the problem wherein the air separation parameters of the conventional air separator cannot be adjusted in a timely manner according to different materials. Based on deep learning, a real-time adjustment system of wind selection parameters was designed. The recognition effects of the four recognition models were tested. Finally, it is found that the comprehensive parameters of the YOLO-AE recognition model are better than the other two models.A real-time wind selection parameter adjustment system was designed based on the previous research on wind selection parameters. The following conclusions were drawn:
The combination of wind separator and deep learning model can improve the quality and efficiency of wind separation. The real-time air separation adjustment system based on deep learning can adjust the air separation parameters in real time with the variation in the proportion of white tea materials. The system identification analysis required less than 2s. Combined with the previous study of winnowing parameters, the adjustment of winnowing machine parameters could be completed within 5s. Through practical experiments, the recognition accuracy and wind selection accuracy of high-quality tea satisfied the requirements of practical application.
The real-time statistics of the proportion of high-quality tea at the discharge outlet could be used to quantitatively analyze the accuracy of tea selection. Compared with the blind parameter adjustment of the conventional air separator, the accuracy of the real-time quantitative analysis of the discharge port can customize the proportion of tea materials. Additionally, it can satisfy the requirements of different types of tea.
Aiming at the requirement of high recognition accuracy for tea wind selection. The YOLO-AE recognition model has better adaptability in the accurate recognition of small target features of white tea materials. This study compares and analyzes the traditional recognition model, although it is similar to YOLO-AE in the number of model parameters. The recognition speed of the improved model is similar to that of the original model, but the accuracy is improved by 3%.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
KL: Methodology, Conceptualization, Validation, Writing – review & editing, Supervision, Formal Analysis, Writing – original draft, Project administration, Funding acquisition, Data curation, Software. YH: Writing – review & editing, Data curation, Conceptualization. XZ: Writing – review & editing, Methodology, Software. ZL: Writing – review & editing, Supervision, Validation. JX: Writing – review & editing, Software, Methodology. DW: Project administration, Methodology, Writing – review & editing, Investigation. RL: Methodology, Software, Writing – review & editing. LT: Software, Validation, Writing – review & editing. YW: Software, Validation, Writing – review & editing, Project administration.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This research was funded by the Natural Science Research Project of Anhui Educational.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Bakshipour, A., Sanaeifar, A., Payman, S. H., and Guardia, M. (2018). Evaluation of data mining strategies for classification of black tea based on image-based features. Food Anal. Methods 11, 1041–1050. doi: 10.1007/s12161-017-1075-z
Chen, Z., Cai, Y., Liu, Y., Liang, Z., Chen, H., Ma, R., et al. (2025). Towards end-to-end rice row detection in paddy fields exploiting two-pathway instance segmentation. Comput. Electron. Agric. 231, 109963. doi: 10.1016/j.compag.2025.109963
Cheng, Y., Li, Y., Zhang, R., Gui, Z., Dong, C., and Ma, R. (2023). Locating tea bud keypoints by keypoint detection method based on convolutional neural network. Sustainability. 15, 6898. doi: 10.3390/su15086898
Dong, C., Wang, M., Chen, Z., Ding, Z., and Tian, Z. (2024). Tea bud detection based on improved YOLOv8 in complex backgrounds. J.Tea Communication 51, 431–440. doi: 10.20178/j.cnki.ISSN1009-525X.2024.04.001
Gan, N., Sun, C. Y., and Zhang, Z. Z. (2018). Research progress of tea processing machinery in China. China Tea Process. 02), 31–37. doi: 10.15905/j.cnki.33-1157/ts.2018.02.011
Hussain, A., Jun, S., Adria, N., Naila, S., Xin, Z., and Kunshan, Y. (2021). Classification of tea varieties based on fluorescence hyperspectral image technology and ABC-SVM algorithm. J. Food Process Preserv. 45, 15241. doi: 10.1111/jfpp.15241
Jiang, M. and Chen, Z. (2020). Symmetry detection algorithm to classify the tea grades using artificial intelligence. Microprocess Microsyst 81, 103738. doi: 10.1016/j.micpro.2020.103738
Liu, Z. H. (2019). The development process and trend of Chinese tea comprehensive processing industry. J. Tea Sci. 39, 115–112. doi: 10.13305/j.cnki.jts.2019.02.001
Luo, L., Cao, C., Wu, Z., Zhang, X., and An, M. (2022). Optimization of roasted green tea winnowing via fluid–solid interaction experiments and simulations. Foods 11, 3271. doi: 10.3390/foods11203271
Lyu, J., Fang, M., Yao, Q., Wu, C., He, Y., Bian, L., et al. (2021). Detection model for tea buds based on region brightness adaptive correction. Trans. Chin. Soc Agric. Eng (Tran CSAE) 37, 278–285. doi: 10.11975/j.issn.1002-6819.2021.22.032
Mingyin, Y., Gangrong, F., Tianbing, C., Muhua, L., Jiang, X., Huamao, Z., et al. (2021). A modified genetic algorithm optimized SVM for rapid classification of tea leaves using laser-induced breakdown spectroscopy. J. Anal. At Spectrom 36, 361–367. doi: 10.1039/D0JA00317D
Mishra, P., Nordon, A., Mohd Asaari, M. S., Lian, G., and Redfern, S. (2019). Fusing spectral and textural information in near-infrared hyperspectral imaging to improve green tea classification modelling. J. Food Eng. 249, 40–47. doi: 10.1016/j.jfoodeng.2019.01.009
Pan, X., Ge, C., Lu, R., Song, S., Chen, G., and Huang, Z. (2022). “On the integration of self-attention and convolution.” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 805–815. doi: 10.1109/CVPR52688.2022.00089
Rahman, M., Ferdous, S., Jenin, M. S., Mim, T., Alam, M., and Mamun, M. (2021). Characterization of tea (Camellia sinensis) granules for quality grading using computer vision system. J. Agric. Food Res. 6, 100210. doi: 10.1016/j.jafr.2021.100210
Rahman, M., Munir, M., and Marculescu, R. (2024). “EMCAD: efficient multi-scale convolutional attention decoding for medical image segmentation.” In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 11769–11779. doi: 10.1109/CVPR52733.2024.01118
Shaohua, H. and Xifeng, L. (2022). Detecting the impurities in tea using an improved YOLOv5 model. Trans. Chin. Soc Agric. Eng (Tran CSAE) 38, 329–336. doi: 10.11975/j.issn.1002-6819.2022.17.036
Sun, J., Tang, K., Wu, X. H., Dai, C., Chen, Y., and Shen, J. (2018). Nondestructive identification of green tea varieties based on hyperspectral imaging technology. J. Food Process Eng. 41, 6. doi: 10.1111/jfpe.12800
Wang, N., Wu, Q., Gui, Y., Hu, Q., and Li, W. (2024). Cross-modal segmentation network for winter wheat mapping in complex terrain using remote-sensing multi-temporal images and DEM data. Remote Sens. 16, 1775. doi: 10.3390/rs16101775
Xu, F., Zhang, K., Zhang, W., Wang, R., Wang, T., Wan, S., et al (2022). Identification and localization method of tea bud leaf picking point based on improved YOLOv4 algorithm. J. Fudan Univ. (Natural Science) 61, 460–471. doi: 10.15943/j.cnki.fdxb-jns.20220728.001
Yan, S., Hanlei, X., Jingming, N., and Zhengzhu, Z. (2018). Grading Keemun black tea based on shape feature parameters of machine vision. Trans. Chin. Soc Agric. Eng (Trans CSAE) 34, 279–286. doi: 10.11975/j.issn.1002-6819.2018.23.036
Yin, C., Su, Y., Pan, M., and Duan, J. (2023). Detection of the quality of famous green tea based on improved YOLOv5s. Trans. Chin/Soc Agric. Eng/(Trans CSAE) 39, 179–187. doi: 10.11975/j.issn.1002-6819.202212151
Yuhan, D., Yuli, Y., Jun, L., Xu, C., and Hui, J. (2022). Classification of tea quality levels using near-infrared spectroscopy based on CLPSO-SVM. Foods 11, 1658–1658. doi: 10.3390/foods11111658
Zhang, X., Yang, T., Qiao, X., Zhing, J., Ou, C. J., Wang, S.-F., et al. (2013). Study on effect of air inlet flow field on the tea particles drift distance and quality of wind-selecting. J. Tea Sci. 33, 60–66. doi: 10.13305/j.cnki.jts.2013.01.006
Zhengmin, W., Yangyang, H., Tengyu, O., Kun, L., Kuan, Q., Yan, S., et al. (2022). Design and test of precise regulation system for process parameters of double-layer tea winnowing machine. Agriculture 12, 1584. doi: 10.3390/agriculture12101584
Keywords: deep learning, tea pneumatic separation, real-time control, YOLO-AE, tea processing
Citation: Luo K, Huang Y, Zhang X, Li Z, Xiong J, Wang D, Liu R, Tao L and Wang Y (2026) Design and experiment of tea winnowing parameter control system based on YOLO-AE. Front. Plant Sci. 16:1721083. doi: 10.3389/fpls.2025.1721083
Received: 03 December 2025; Accepted: 22 December 2025; Revised: 24 November 2025;
Published: 21 January 2026.
Edited by:
Yu Nishizawa, Kagoshima University, JapanReviewed by:
Nitin Goyal, Central University of Haryana, IndiaMatthew Oladipupo, University of Salford, United Kingdom
Copyright © 2026 Luo, Huang, Zhang, Li, Xiong, Wang, Liu, Tao and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yujie Wang, d2FuZ3lqbmV0QDE2My5jb20=
†These authors have contributed equally to this work
Yangyang Huang2†