- 1School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, China
- 2Institute of Characteristic Agriculture, Jiaozuo Academy of Agriculture and Forestry Sciences, Jiaozuo, China
Introduction: Yam is an important medicinal and edible crop, but its quality and yield are greatly affected by leaf diseases. Currently, research on yam leaf disease segmentation remains unexplored. Challenges like leaf overlapping, uneven lighting and irregular disease spots in complex environments limit segmentation accuracy.
Methods: To address these challenges, this paper introduces the first yam leaf disease segmentation dataset and proposes BiSeNeXt, an enhanced method based on BiSeNetV2. Firstly, dynamic feature extraction block (DFEB) enhances the precision of leaf and disease edge pixels and reduces lesion omission through dynamic receptive-field convolution (DRFConv) and pixel shuffle (PixelShuffle) downsampling. Secondly, efficient asymmetric multi-scale attention (EAMA) effectively alleviates the problem of lesion adhesion by combining asymmetric convolution with a multi-scale parallel structure. Finally, PointRefine decoder adaptively selects uncertain points in the image predictions and refines them point-by-point, producing accurate segmentation of leaves and spots.
Results: Experimental results indicated that the approach achieved a 97.04% intersection over union (IoU) for leaf segmentation and an 84.75% IoU for disease segmentation. Compared to DeepLabV3+, the proposed method improves the IoU of leaf and disease segmentation by 2.22% and 5.58%, respectively. Additionally, the FLOPs and total number of parameters of the proposed method require only 11.81% and 7.81% of DeepLabV3+, respectively.
Discussion: Therefore, the proposed method can efficiently and accurately extract yam leaf spots in complex scenes, providing a solid foundation for analyzing yam leaves and diseases.
1 Introduction
Yam (Dioscorea spp.) is a tuberous plant of the genus Dioscorea within the family Dioscoreaceae (Diouf et al., 2022; Baffour-Ata et al., 2023). Yam is rich in starch, proteins, minerals, polysaccharides, saponins, flavonoids and other biologically active substances (Gao et al., 2023a; Li et al., 2024). It is of great value in the field of food and medicine (Li et al., 2023). China is one of the major producers of yam (Chen et al., 2022). It is primarily cultivated in Henan, Hebei, Shandong, Jiangxi and Yunnan provinces (Gao et al., 2023b). However, various diseases frequently occur in the cultivation of yams, with leaf diseases being particularly common. These diseases are mainly caused by viral, bacterial and fungal infections (Gogile et al., 2024). Such infections can significantly impair leaf photosynthesis, ultimately affecting yam growth and yield (Zhao et al., 2022; He et al., 2025). Currently, deep learning-based research on crop leaf disease analysis primarily focuses on crops such as tomatoes, apples and grapes, while studies on yam leaf disease analysis remain absent. This study proposes a solution for the segmentation of yam leaf diseases for the first time, providing technical support for yam disease identification and precision planting. Taking anthracnose as an example, the infection rate reaches 92.05% (Wang et al., 2024), posing a significant threat to yam growth and yield, with losses ranging from 50% to 90% (Tariq et al., 2024). Therefore, accurate analysis and targeted treatment of yam leaf diseases are essential for ensuring healthy yam growth. Computer vision methods provide an effective solution by helping growers identify affected areas and apply medication precisely (Tariq et al., 2024; Lu et al., 2023).
Image segmentation algorithms are widely used to analyze leaves and diseases, with segmentation results enabling the calculation of leaf and disease spot areas to assess disease severity. Conventional segmentation approaches are usually divided into three types: (1) Region-based methods. For example, Kumar and Sachar (2023) employed integrated color features combined with region-growing techniques and multi-resolution channels to segment diseased areas of crops. (2) Clustering-based methods. For example, Ronneberger et al. (2015) applied the K-means clustering method to segment cotton and tomato leaf images, and Long et al. (2015) utilized fuzzy C-means (FCM) combined with the chameleon clustering algorithm to segment diseased portions of plant leaves. (3) Thresholding-based methods. For example, Badrinarayanan et al. (2017) employed a color vegetation index combined with the Otsu thresholding segmentation method to achieve accurate segmentation of diseased leaves in cruciferous crops. Traditional approaches mainly focus on local pixel relationships, which may result in local optimal solutions. The effectiveness of these methods is significantly diminished when the diseased region of the crop exhibits a similar hue to the background or has blurred boundaries (Kumar and Sachar, 2023). Moreover, these methods often require manual parameter adjustments, offering limited automation, which makes them unsuitable for large-scale applications. Furthermore, traditional methods are generally designed for specific diseases and lack universality across different types of disease, limiting their broad practical applicability.
Recently, artificial intelligence has seen rapid advancements. The fully convolutional network (FCN) (Long et al., 2015) initially employed convolutional neural networks for semantic segmentation at the pixel-level. Subsequently, CNN-based architectures including U-Net (Ronneberger et al., 2015), various versions of DeepLab (Chen, 2014; Chen et al., 2017; Chen, 2017), PSPNet (Zhao et al., 2017) and SegNet (Badrinarayanan et al., 2017) have emerged. Their high accuracy and transferability have drawn considerable interest from the research community. These models have been widely used for crop disease recognition (Picon et al., 2022; Zhang et al., 2023c), achieving better results compared to traditional image segmentation algorithms (Liu et al., 2024). These methods perform well in addressing segmentation tasks for various diseases, particularly in simpler environments. However, in outdoor environments, the accuracy of segmentation results decreases due to factors such as leaf overlap, leaf curling, lighting variations and water droplet interference. To address these issues, researchers have proposed a variety of structurally complex models to improve the segmentation performance of leaf disease images. Zhang and Zhang (2023) proposed an improved U-Net model, MU-Net, which incorporates residual blocks (ResBlock) and residual paths (ResPath) to enhance feature learning. This design effectively improves the segmentation performance of plant disease leaf images. Zhou et al. (2024) proposed a two-stage 3D point cloud-based method for orchid phenotypic analysis, significantly improving the efficiency and accuracy of phenotypic parameter measurement in orchid seedlings. However, the method may face challenges when dealing with diverse disease morphologies and complex environmental conditions. Yang et al. (2025) proposed a segmentation model named FATDNet, which incorporates adversarial networks to enhance performance. By introducing a dual-path fusion adversarial algorithm (DFAA), a multi-dimensional attention mechanism (MDAM), and a gaussian weighted edge segmentation module (GWESM), the model significantly improves the segmentation accuracy of tomato leaf diseases. To improve the segmentation accuracy and feature extraction capabilities of small spots on apple leaves, Zhang et al. (2023a) proposed an improved DFL-UNet+CBAM model that combines a hybrid loss function of Dice Loss and Focal Loss with the convolutional block attention module (CBAM). However, these methods generally suffer from high computational complexity and large numbers of parameters, limiting their potential for practical deployment.
To reduce computational resource demands and enhance the feasibility of model deployment in real-world scenarios, researchers have increasingly focused on developing lightweight methods for leaf disease segmentation. Wang et al. (2025) proposes a lightweight decoupled saliency detection method with excellent boundary refinement. However, the method was primarily designed for binary classification tasks, and its applicability to multi-label leaf disease segmentation remains limited. Zhu et al. (2023) proposed a lightweight two-stage LD-DeepLabv3+ model that uses receptive field blocks (RFB), reverse attention (RA), and channel attention block (CAB) to improve feature extraction. With adaptive loss to handle lesion pixel imbalance, it enhances segmentation accuracy in complex environments. However, the use of dilated convolutions in the decoder may be insufficient for recovering spatial details, which can easily lead to the omission or adhesion of small lesions. Shwetha et al. (2024) proposed a lightweight U-Net model based on MobileNetV4, combined with data augmentation using a deep convolutional generative adversarial network (DCGAN), to improve the segmentation accuracy of small targets in complex backgrounds caused by jasmine leaf spot disease. However, the use of transposed convolution in the decoder may introduce grid-like artifacts due to non-uniform pixel interpolation, leading to a decline in the recovery of details such as lesion boundaries. Dai et al. (2024) proposed the AISOA-SSFormer model, which integrates a sparse global update perceptron (SGUP) and an annealing integrated sparrow optimization algorithm (AISOA) to improve the accuracy of rice leaf disease identification in complex scenarios. However, the random perturbations in the sparrow search algorithm may cause fluctuations in the loss function, requiring more iterations and potentially losing part of the boundary information. Table 1 presents an overview of methods used for identifying plant leaf diseases. These lightweight leaf disease segmentation methods hold strong potential for deployment on mobile devices, but may lack precision in leaf disease segmentation. However, the reliance on complex models and high computational demands makes it difficult to achieve both high accuracy and computational efficiency.
Existing methods have demonstrated strong performance in segmenting leaf disease images with single backgrounds and high resolution, and some have also shown progress in more complex environments. However, these methods often rely on complex architectures and high computational resources, making it challenging to balance segmentation accuracy with computational efficiency. To address these challenges, this paper presents the BiSeNeXt network, designed to achieve accurate segmentation with reduced model complexity. The proposed method has been tested on yam leaf disease images, demonstrating its efficiency and reliability for health analysis. The main contributions of this paper are as follows:
1. In this paper, we construct the first high-quality leaf disease dataset of yam, covering three major diseases: anthracnose, brown spot and gray spot. The dataset consists of images captured in various weather conditions, including sunny, rainy and cloudy days.
2. To improve the segmentation accuracy of leaf and spot boundaries and reduce spot omission, we propose dynamic feature extraction block (DFEB). This block integrates dynamic receptive-field convolution (DRFConv) to enhance boundary and complex structure modeling capabilities. Meanwhile, it uses pixel shuffle (PixelShuffle) downsampling to effectively alleviate information loss and improve detail retention.
3. To address the issue of adhesion of adjacent lesions and reduce misclassification, we propose the EAMA attention. It combines asymmetric convolutions with a multi-scale parallel structure to effectively capture complex spatial features and multi-scale contextual information.
4. To effectively recover details and edge information during downsampling, we propose the PointRefine decoder. By adaptively selecting uncertain key points in the image for point-by-point refinement, this decoder significantly enhances the restoration of fine details and edges.
The rest of the paper is structured as follows: Section 2 introduces the dataset, preprocessing steps and details of the proposed BiSeNeXt model architecture. Section 3 explains the experimental setup, evaluation metrics and result analysis. Finally, Section 4 concludes the study.
2 Materials and methods
2.1 Data acquisition
Due to the absence of publicly available yam leaf disease datasets, this study independently constructed a high-quality dataset including three major yam leaf diseases. The dataset was collected from April to September 2024 at Jiaozuo Academy of Agricultural and Forestry Sciences and in Xiaoyou Village, Jiaozuo, Henan, China. The dataset consists of 1,097 high-resolution images, including 396 images of anthracnose, 363 images of brown spot and 338 images of gray spot, as is shown in Table 2. The name “gray spot” was determined through discussions with multiple agricultural experts to ensure its scientific validity and accuracy. Specifically, indoor images were taken using a Canon EOS 70D on a tripod at a distance of about 35 cm, saved in JPEG format with a resolution of 5472 × 3648 pixels, while outdoor images were captured handheld with an iPhone XR at a distance of 25–45 cm, saved as JPG files with a resolution of 3072 × 3072 pixels. Images were collected under sunny, cloudy and rainy conditions. The majority were captured on well-lit sunny days, while a smaller portion was taken on rainy days to enhance the dataset’s diversity and comprehensiveness. The dataset includes natural variations such as leaf overlap, occlusions and varied shooting angles, deliberately incorporated to reflect real-world field conditions. Figure 1 illustrates a typical scene from a yam field, depicting the actual environment used for capturing disease images.
Figure 2 presents representative images of three yam leaf diseases, illustrating their distinct characteristics and impact on the plant. Anthracnose primarily infects the leaves, petioles, stems and vines of yam. In the early stages, the lesions appear as small, round, dark brown, water-soaked spots. As the disease progresses, the lesions enlarge into large brown or dark brown patches, some forming concentric rings. In severe infections, the lesions cause leaf margin desiccation and shedding, ultimately affecting yam yield and quality. Brown spot is a common yam leaf disease, primarily characterized by irregular yellow to brown spots that gradually darken to deep brown, ultimately resulting in leaf necrosis. This disease impairs plant photosynthesis and growth, consequently reducing yield and impacting the quality and economic value of yams. Gray spot primarily infects the leaves. Initially, it forms small yellowish spots that gradually expand into oval brown lesions. Ultimately, it causes premature leaf senescence and shedding, thereby further impacting yam growth and yield.

Figure 2. Images and labels of three yam leaf diseases. In the annotations, red, green and black represent lesions, leaves, and background, respectively.
As illustrated in Figure 2, yam leaf disease segmentation requires accurately identifying both leaf structures and diseased regions, which presents several challenges. Leaf segmentation faces several challenges: (1) Leaf overlap significantly complicates edge extraction in real-world outdoor environments. (2) Shadows from uneven illumination pose challenges in capturing the global features of the leaf. (3) Leaf curling causes changes in contour shape, interfering with accurate boundary segmentation. Disease segmentation also presents several challenges: (1) Raindrops on the leaf surface and uneven lighting conditions complicate the segmentation of disease spots. (2) Disease spots are typically small, densely clustered and randomly distributed. This characteristic not only increases the risk of missed detections but also causes spot adhesion, thereby reducing segmentation accuracy. (3) Irregular lesion shapes and blurred boundaries pose significant challenges for lesion segmentation.
2.2 Data processing
During data preprocessing, the original images were initially annotated. To ensure annotation accuracy and precision, the professional semantic segmentation tool Labelme was used for pixel-level fine annotation under expert guidance. During annotation, the edges and diseased areas of each leaf were accurately delineated, and the corresponding semantic segmentation label maps were generated, as shown in Figure 2. Subsequently, all images and their corresponding annotations were resized to a fixed resolution of 512×512 to meet the model’s input requirements. The high-quality pixel-level annotations provide a robust dataset for training and evaluating the subsequent segmentation model.
2.3 Data augmentation
Neural networks require ample training data, as insufficient samples can lead to overfitting and hinder generalization (Lee et al., 2019). Therefore, it is necessary to expand the original yam leaf disease dataset appropriately. Figure 3 shows a yam anthracnose as an example. Six data augmentation methods are applied to enhance the image, including adjusting saturation, brightness and contrast, as well as cropping, rotation and flipping. These data augmentation methods effectively enhance the model’s generalization and adaptability to different environments by simulating scenes with varying camera angles and lighting conditions. Table 2 provides details on the specific number of images generated from indoor and outdoor photography, as well as the augmentation applied to the three types of yam leaf diseases.

Figure 3. Image enhancement. (A) Original image. (B) Saturation change. (C) Brightness change. (D) Contrast enhancement. (E) Image crop. (F) Image rotation. (G) Image flip.
2.4 The proposed method
This section introduces BiSeNeXt, an improved deep learning model. It aims to improve the segmentation accuracy of yam leaf diseases in complex environments. Traditional deep learning segmentation methods often perform poorly when encountering challenges such as leaf overlap, light variation, raindrop interference and small lesion areas. To address these issues, the BiSeNeXt network was optimized based on BiSeNetV2 (Yu et al., 2021) and primarily consists of DFEB, EAMA and PointRefine modules. The design of these modules enhances the network’s ability to segment details and lesion regions in complex backgrounds.
2.4.1 Overall structure of BiSeNeXt
The architecture of the BiSeNeXt network is illustrated in Figure 4. The green dashed box highlights the two backbone networks. The detail branch at the top extracts fine-grained spatial features, while the semantic branch at the bottom captures high-level contextual information. The numbers inside the cubes represent feature map resolution relative to the input. The yellow dashed box represents the Aggregation Layer, where “Down” and “Up” indicate downsampling and upsampling operations, respectively. Meanwhile, the Sigmoid activation function performs element-wise multiplication. The symbol ⊗ denotes the element-wise multiplication operation and φ represents the Sigmoid function. In addition, the booster component, highlighted by the blue dashed box, enhances segmentation performance during training using multiple auxiliary segmentation heads (Seg Head). This process does not increase the inference cost.
To visualize the structural configuration of detail branch and semantic branch, Table 3 presents the specific operations and their corresponding parameters for each stage. Each stage S comprises one or more operations (opr), including DFEB, EAMA, Stem, gather-and-expansion layer (GE) and context embedding block (CE), where c represents the number of output channels, s denotes the stride, r indicates the number of repetitions, and e signifies the channel expansion factor.
2.4.2 Dynamic feature extraction block
The complex and diverse shapes of lesions in yam leaf disease images pose significant challenges for feature extraction, often resulting in missed spot regions and reduced segmentation performance. In convolutional neural networks, standard convolutions extract local features by sharing parameters across sliding windows. However, traditional convolutions struggle to capture spatial differences effectively because they use shared weights across all positions within the sliding window. Additionally, traditional downsampling methods often lead to information loss, especially in fine details and boundary regions, which negatively impacts segmentation accuracy. Inspired by receptive-field attention convolutional (RFAConv) (Zhang et al., 2023b), this paper proposes the dynamic feature extraction block (DFEB), which integrates dynamic receptive-field convolution (DRFConv) with a downsampling method based on pixel shuffle (PixelShuffle). The DFEB module mitigates lesion omission by employing efficient downsampling and detail preservation techniques, It enhances both feature extraction and segmentation performance for complex lesion morphologies.
DRFConv uses dynamically generated attention weights to assign unique parameters to the convolution kernel. This resolves the limitations of parameter sharing in standard convolution. It also improves the network’s capacity to capture complex spatial patterns. As shown in Equation 1, attention weights are computed using the Softmax activation function. They are applied element-wise to the convolution kernel parameters, emphasizing the significance of different features.
In Equation 1, X denotes the input feature map, and F represents the augmented features generated by DRFConv. Here, stands for grouped convolution with a kernel size of n × n, k indicates the size of the convolution kernel, and Norm refers to normalization.
Figure 5 illustrates the detailed structure and computational flow of DRFConv. The input feature map is of size C × H × W, with C indicating the channel count, H representing the height and W denoting the width. Initially, group convolution extracts spatial features from the receptive-field, forming non-overlapping sliders. This process produces a feature map of size 9C × H × W and mitigates the high computational burden of the traditional unfold operation. Next, the global information of each slider is aggregated through global average pooling, and feature interactions are facilitated via 1 × 1 group convolution. Attention weights are generated by applying the Softmax function, further emphasizing the importance of different features. The resulting feature map of size 9C ×H ×W is multiplied by the feature map generated by group convolution. Finally, the “Adjust Shape” operation modifies the feature map’s dimensions, multiplying both its height and width by a factor of k. Feature information is then extracted using a k × k convolution operation with stride k. This design effectively addresses the parameter-sharing issue in standard convolution, enhancing the network’s ability to model complex spatial features while maintaining computational efficiency.
In addition, PixelShuffle mitigates the issue of information loss in traditional downsampling by employing tensor rearrangement. This not only addresses feature loss in conventional methods, but also enhances computational efficiency and improves boundary processing accuracy. PixelShuffle addresses information loss in traditional downsampling by splitting the input feature maps into multiple subregions and rearranging them into higher-dimensional channels. This results in efficient reorganization of features. Compared to pooling downsampling and convolutional downsampling, PixelShuffle does not need to discard input features. Instead, it completes the resolution reduction through information rearrangement, which reduces information loss. Moreover, the method is computationally efficient, avoids additional weight learning, and excels in detail preservation and boundary processing. This enhances its applicability to fine segmentation in image processing. Figure 6B presents the core structure of the module.

Figure 6. The structure of the DFEB module. (A) DFEB structure with a stride of 1. (B) DFEB structure with a stride of 2.
By integrating DRFConv and PixelShuffle, DFEB enhances the network’s capability to model spatial features while improving detail and boundary processing. Figure 6 illustrates the structure of DFEB. Figure 6A represents the DFEB structure with a stride of 1. The procedure is as follows: First, a 3 × 3 DRFConv is applied to modify the number of channels and capture both global and local features. Next, a 3 × 3 convolution is applied to emphasize the details of the edges and the texture characteristics in the leaf and disease regions. Finally, a 1 × 1 convolution serves as a shortcut connection to adjust the number of channels in the feature map and enhance model training stability. Figure 6B illustrates the DFEB structure with a stride of 2. Assuming the input feature map has dimensions C × H × W, the main branch is initially downsampled using a 3 × 3 DRFConv, followed by local feature extraction through a 3 × 3 convolution. The side branch is downsampled using PixelShuffle. PixelShuffle first applies necessary padding to the right and bottom of the input features to ensure that their height and width are even. Next, PixelShuffle is applied, increasing the number of channels to λ2C. The data is then rearranged to distribute channel information across the spatial dimensions with height and width , yielding feature maps of size . A 1 × 1 convolution is then applied to adjust the number of channels, and the result is finally summed and fused with the main branch features. Notation: BN denotes batch normalization. ReLU refers to the ReLU activation function. Conv represents the convolution operation. λ represents the downsampling factor, which is set to 2 in this study.
2.4.3 Efficient asymmetric multi-scale attention
When segmenting dense and small lesions in yam leaf images, adhesion between lesions and overlapping leaves often leads to mis-segmentation, making accurate identification challenging. Existing attention mechanisms, such as the Convolutional Block Attention Module (CBAM) (Woo et al., 2018) and Squeeze-and-Excitation (SE) (Hu et al., 2018), typically reduce computational complexity through channel dimension compression. However, this approach often results in the loss of fine-grained spatial and channel information, thereby limiting deep feature representation capabilities. In addition, existing methods struggle to model feature scales effectively, making it challenging to refine local details while preserving global context. To address these issues, this paper proposes an improved efficient asymmetric multi-scale attention (EAMA). It is based on the spatial coding strategy of the efficient multi-scale attention (EMA) (Ouyang et al., 2023) and the concept of parallel multi-scale convolutions. The module combines asymmetric convolution with a parallel modeling structure. This design aims to enhance multi-scale feature extraction while balancing computational efficiency and accuracy. By constructing parallel substructures at multiple scales, EAMA alleviates the need for deep network hierarchies and sequential processing, thereby improving the model’s capacity for multi-scale representation. Asymmetric convolution enhances the model’s ability to perceive structural variations in different directions while maintaining low computational cost. Part of the channel dimension is restructured into the batch dimension, enabling cross-channel interaction without channel compression. This approach reduces computational complexity while preserving fine-grained features. EAMA further introduces a local cross-spatial and cross-channel collaborative modeling strategy to enhance feature fusion and the integration of contextual information. By combining adaptive weighting with multi-scale feature fusion, EAMA significantly enhances the representation of complex structural features while improving computational efficiency. This makes it particularly effective in challenging segmentation scenarios such as lesion adhesion and leaf overlap.
As illustrated in Figure 7, the EAMA module initially groups the input feature map by channel dimensions, dividing the channels into G subgroups, each containing channels. The grouped features are then reshaped to dimensions , which enables independent modeling of spatial and channel relationships within each feature group. To minimize sequential processing and excessive network depth, EAMA uses a parallel substructure with a 1×1 branch and an asymmetric branch. The 1 × 1 branch consists of two paths, each applying one-dimensional global average pooling along the horizontal and vertical dimensions, respectively, as defined in Equation 2.
Here, xc is the input feature map, with c indicating the number of input channels, while H and W correspond to the spatial dimensions of the input features. and represent the global information of the channel c along the height and width directions, respectively. Next, two-channel vectors are generated by decomposing the fused features through a channel splicing operation using a shared 1 × 1 convolution. Subsequently, two-channel attention maps are computed using a nonlinear Sigmoid function. To capture distinct cross-channel interaction features between the two parallel routes of the 1 × 1 branch, the two-channel attention maps are aggregated via element-wise multiplication within each group. Meanwhile, the asymmetric convolutional branch processes feature maps in parallel using 1 × 3, 3 × 1 and 3 × 3 convolutions, then sums them to capture multi-scale features and enhance spatial feature representation. After obtaining the outputs from the 1 × 1 branch and asymmetric branch, the EAMA module further fuses the features using a cross-space learning strategy. EAMA applies 2D global average pooling to the outputs of the 1 × 1 branch to encode global spatial information, as defined in Equation 3:
This design encodes global spatial information. The first spatial attention map is subsequently generated through normalization using the Softmax function, followed by a dot product computation with the asymmetric branch’s output. Another branch uses 2D global average pooling to capture global spatial information. This is normalized via the Softmax function. A dot product is then computed with the 1 × 1 branch’s output, producing the second spatial attention map. This map captures multi-scale spatial information. The Sigmoid function processes the sum of the two spatial attention maps, producing the final attention map that adjusts pixel weights in the input feature map.
2.4.4 PointRefine
Traditional convolutional neural networks often suffer from oversampling in smooth regions and undersampling near object boundaries during leaf disease image segmentation. To address these challenges, this paper proposes a novel decoder named PointRefine, inspired by PointRend (Kirillov et al., 2020). PointRefine introduces a non-uniform sampling mechanism to perform point-by-point prediction on key areas with high-frequency features in the output image. This method enables more accurate feature reconstruction in boundary regions while reducing redundant computation in smooth areas, effectively balancing segmentation accuracy and efficiency. Furthermore, PointRefine effectively restores features extracted by the encoder, improving the model’s ability to distinguish difficult-to-classify pixels. The structure of PointRefine is shown in Figure 8.
PointRefine generates high-quality segmentation results through a point-by-point refinement process. PointRefine uses the Seg Head to generate a low-resolution coarse prediction, which is then upsampled via bilinear interpolation. To further refine the segmentation boundary, j × N points are sampled from the coarse prediction. The βN (β ∈ [0,1]) most uncertain points among them are selected as key points. Uncertainty is determined by the absolute difference between the logits of the top two categories in the coarse prediction, where smaller differences indicate higher uncertainty. These key points are typically located at object boundaries or complex regions. To ensure comprehensive region coverage, the remaining (1 − β) × N points are obtained via random sampling with a uniform distribution. Point features are processed by a fully connected layer, after which they are concatenated with coarse predictions and input into the multilayer perceptron (MLP). Each point’s label is independently predicted by the MLP, with new predictions iteratively updated in the feature map. During inference, the refinement process is repeated m times, producing high-resolution segmentation with clear boundaries, as shown in Figure 8. In this research, the parameters are set as follows: j = 3, N = 2048, β = 0.75 and m = 2.
3 Experiments
This section provides a detailed description of the experimental setup, evaluation metrics and experimental results. Through ablation experiments, we systematically analyze the impact of key model components on performance. In addition, through comparative experiments, we comprehensively evaluate the advantages and limitations of the proposed method.
3.1 Experimental setup
The experimental environment is configured with an Intel® Xeon® Gold 6330 CPU (2.00 GHz), a GeForce RTX 3090 GPU (24 GB graphics memory) and a 64-bit Linux operating system. The model is implemented using the PyTorch deep learning framework. The software environment consists of PyTorch 1.12.1 and Python 3.9. The training hyperparameters are set as follows: an initial learning rate of 1e-2, a minimum learning rate of 1e-4, a batch size of 8, 200 epochs, a momentum of 0.9, a weight decay of 5e-4 and optimization using the SGD optimizer. A learning rate decay strategy is applied to facilitate stable model convergence. The training time per epoch is approximately 209 seconds.
The custom yam dataset used in this study contains images with a resolution of 512 × 512 × 3, and is divided into training, validation, and testing sets in a 7:2:1 ratio. Each subset comprises various yam leaf disease types. This enables the model to learn features across different classes and enhances its generalization and robustness.
3.2 Evaluation indicators
To evaluate the performance of the yam leaf disease segmentation network in complex environments, various metrics are used, including intersection over union (IoU), mean IoU (mIoU), Precision, Recall and Dice coefficient (Dice) and mean F1-Score (mF1-Score). In addition, inference time (Inf. time), total parameters and floating point operations (FLOPs) are used to assess the model’s efficiency and computational complexity.
In semantic segmentation tasks, IoU measures accuracy by calculating the ratio of the intersection region to the union region between predicted and true labels. mIoU represents the average IoU across all categories and is used to assess overall segmentation performance in multi-category tasks. The mF1-Score is used to measure the average segmentation performance of the model across all categories to avoid bias due to category imbalance. Precision measures the proportion of correctly predicted positive pixels among all pixels predicted as positive, reflecting the model’s ability to avoid false positives. Recall measures the proportion of correctly predicted positive pixels among all actual positive pixels, indicating the model’s ability to detect relevant regions in segmentation tasks. Dice measures segmentation accuracy by calculating twice the intersection over the sum of predicted and true positive regions. IoU, mIoU, Precision, Recall, Dice and mF1-Score are defined in Equations 4–9:
Here, k represents the number of target classes considered in the computation after excluding background classes, with k = 2 in this experiment. pij denotes the number of pixels belonging to class i but misclassified as class j. TP is the number of samples correctly identified as positive. FP is the number of negative samples incorrectly predicted as positive. FN is the number of positive samples incorrectly predicted as negative.
3.3 Effect of data augmentation on model performance
To assess the impact of data augmentation on the performance of the baseline model, we conducted comparative experiments with and without data augmentation. Table 4 presents the IoU scores for each category and the overall mF1-Score. The results show that the model with data augmentation outperforms the baseline model without augmentation across all metrics. Specifically, the IoU scores for the background, leaf and disease categories improved by 2.09%, 4.06% and 4.75%, respectively. The overall mF1-Score increased by 2.06%. These enhancements indicate that data augmentation significantly improves the model’s generalization ability. It also substantially improves the model’s segmentation accuracy and robustness in segmenting the target region.
3.4 Comparison of different attention mechanisms
The detail branch of BiSeNetV2 is designed to capture high-resolution spatial details. However, it has limitations in handling the adhesion of adjacent disease spots on yam leaves. To improve the detail branch’s ability to model the texture and boundaries of leaf disease regions, this study incorporated multiple attention mechanisms into BiSeNetV2. Table 5 presents the performance changes following the integration of five different attention mechanisms into BiSeNetV2. The experimental results indicate that efficient channel attention (ECA) (Wang et al., 2020), CBAM (Woo et al., 2018), SE (Hu et al., 2018), enhanced parallel attention (EPA) (Lu et al., 2024) and EAMA exhibited changes of 0.04%, -0.15%, 0.02%, -0.15% and 0.23% in the IoU for yam leaf segmentation; 0.15%, -0.43%, -0.13%, 0.11% and 0.63% in the IoU for yam disease segmentation; and 0.05%, -0.05%, 0.02%, -0.11% and 0.19% in the mF1-Score. EAMA achieved the best performance across all metrics. It shows outstanding capability in modeling complex spatial relationships and disease segmentation.
3.5 Discussion of various loss functions
In this section, the effects of different loss functions on the segmentation performance of BiSeNetV2 are explored through comparative experiments. As shown in Table 6, CE Loss improves IoU in leaf segmentation by 2.37%, 0.59% and 0.08% compared to Focal Loss, Dice Loss and Ohem+CE Loss, respectively. For disease segmentation, it achieves improvements of 7.43%, 2.62% and 0.30%, respectively. In addition, CE Loss also significantly outperforms other loss functions on mF1-Score by 2.26%, 0.68% and 0.10%, respectively. The experimental results show that CE Loss performs superiorly in all indicators, especially in disease segmentation accuracy, which is significantly improved. Based on these results, CE Loss was ultimately selected as the loss function for the model in this study.
3.6 Ablation experiments
This subsection presents eight sets of ablation experiments to evaluate the effectiveness of the DFEB, EAMA and PointRefine modules in BiSeNetV2. These experiments focus on improving leaf and disease segmentation performance using the controlled variable method. The optimal results for each metric are highlighted in bold and a ✓ symbol indicates the inclusion of the corresponding module. The experimental results are presented in Table 7.
As shown in Table 7, in single-module ablation experiments, Test 1 corresponds to the baseline BiSeNetV2 model. By adding the PointRefine, EAMA and DFEB modules to the baseline, the experimental results show that the introduction of each module improves the model performance. The results of Test 2 demonstrate that the introduction of the PointRefine module improves leaf IoU, disease IoU, mF1-Score and mIoU by 0.63%, 1.62%, 0.48% and 0.83%, respectively. In addition, the number of parameters is reduced by 1.02M, and FLOPs decrease by 4.04G. This indicates that PointRefine enhances the model’s ability to distinguish hard-to-classify pixels, thereby producing more accurate image segmentation results. The results of Test 3 show that the introduction of the EAMA module improves leaf IoU, disease IoU, mF1-Score and mIoU by 0.23%, 0.63%, 0.19% and 0.32%, respectively. This validates the effectiveness of the EAMA module in extracting multi-scale features, which significantly improves the segmentation performance of leaves and disease spots, while keeping the number of parameters and FLOPs nearly unchanged. In Test 4, the DFEB module improves leaf IoU, disease IoU, mF1-Score and mIoU by 0.28%, 0.49%, 0.38% and 0.30%, respectively. The DFEB module improves leaf disease boundary segmentation, but increases parameters and FLOPs by 0.91 M and 12.19 G, respectively.
In multi-module ablation experiments, we evaluate the effects of the DFEB, EAMA and PointRefine modules on model performance through various combinations. As shown in Table 7, the module combination performs better than adding modules individually on multiple indicators. It fully demonstrates the combined effect between modules. In Test 8, the simultaneous combination of the three modules improves the leaf IoU, disease IoU, mF1-Score and mIoU by 0.98%, 3.96%, 1.02% and 1.76%, respectively. Meanwhile, the number of model parameters decreases by 0.12 M, while FLOPs increase by 8.53 G. The improved BiSeNeXt network achieves optimal performance in the IoU, mF1-Score and mIoU metrics. Considering the balance between segmentation performance and computational efficiency, the improved model effectively meets the requirements of the yam leaf disease segmentation task.
Figure 9 presents the visualization results of different models for segmenting diseased regions of yam leaves in the ablation experiments. The baseline model Figure 9C exhibits noticeable omissions and misdetections when segmenting diseased regions with complex boundaries or fine shapes. In addition, severe adhesion occurs in some adjacent disease regions, where multiple separate disease lesions are erroneously merged into one region. Comparison among Figures 9B, C, E–G reveals that after integrating individual modules, each module demonstrates its respective advantages in addressing segmentation challenges. PointRefine has clear advantages in the restoration of fine details and edges. EAMA enhances multi-scale features to improve the model’s segmentation of large diseased regions and effectively reduces spot adhesion. DFEB enhances boundary processing in complex regions through dynamic receptive-fields and boundary optimization. Further comparison of the plots in Figures 9B–D, H–J shows that combining multiple modules improves model performance in complex boundary processing, fine-grained disease segmentation, and mitigation of the spot adhesion phenomenon. When all three modules are optimized jointly Figure 9J, the model achieves its best segmentation performance, significantly surpassing the baseline model Figure 9C. This fully validates the effectiveness and rationality of multi-module joint optimization.

Figure 9. Results of ablation experiments. (A) Original images. (B) Ground truth. (C) BiSeNetV2. (D) BiSeNeXt(Ours). (E) BiSeNetV2+DFEB. (F) BiSeNetV2+EAMA. (G) BiSeNetV2+PointRefine. (H) BiSeNetV2+DFEB+EAMA. (I) BiSeNetV2+DFEB+PointRefine. (J) BiSeNetV2+EAMA+PointRefine.
3.7 Comparison of different segmentation networks
To evaluate the effectiveness of the proposed network in segmenting yam leaves and diseases in complex surroundings, we compare it with other popular semantic segmentation models. This paper selected DeepLabV3+ (Chen et al., 2018), DFL-UNet+CBAM (Zhang et al., 2023a), UNet++ (Zhou et al., 2018), UNet (Ronneberger et al., 2015) and PSPNet (Zhao et al., 2017) as representative CNN-based comparison methods, and Segmenter (Strudel et al., 2021), Swin Transformer (Liu et al., 2021), Vision Transformer (Dosovitskiy et al., 2020) and SegNeXt (Guo et al., 2022) as representative Transformer-based comparison methods. To maintain fair conditions, all models were trained and tested on the same custom yam dataset. Table 8 shows the segmentation performance of different models.
As presented in Table 8, the proposed method achieved an IoU of 84.75% and a recall of 91.74% for disease segmentation. The BiSeNeXt method proposed in this study demonstrates superior performance compared to mainstream segmentation models on the yam leaf test set, including DeepLabV3+, DFL-UNet+CBAM, UNet++, U-Net, PSPNet, Segmenter, SegNeXt, Swin Transformer and Vision Transformer. Specifically, it increased the IoU for leaf segmentation by 2.22%, 3.22%, 1.09%, 1.22%, 1.56%, 19.60%, 4.84%, 5.39% and 3.83%, respectively. Additionally, the IoU for disease segmentation improved by 5.58%, 3.98%, 0.75%, 0.76%, 5.01%, 26.21%, 9.71%, 6.86% and 6.20%, respectively. The precision of leaf and disease segmentation improved by 1.46%, 2.34%, 1.12%, 1.12%, 1.01%, 15.90%, 4.25%, 4.86%, 3.16%, and 3.79%, 0.75%, 0.24%, 0.22%, 2.03%, 9.00%, 2.45%, 1.81%, 3.02%. The recall for leaf and disease segmentation improved by 0.84%, 1.01%, 0, 0.13%, 0.60%, 5.92%, 0.79%, 0.74%, 0.83%, and 2.95%, 3.95%, 0.64%, 0.67%, 3.99%, 25.07%, 9.28%, 6.41%, 4.49%. The Dice for leaf and disease segmentation increased by 1.16%, 1.69%, 0.57%, 0.63%, 0.81%, 11.21%, 2.56%, 2.85% 2.02% and 3.37%, 2.37%, 0.44%, 0.44%, 3.01%, 17.89%, 6.00%, 4.17%, 3.76%. The mIoU increased by 2.90%, 2.88%, 0.78%, 0.85%, 2.39%, 18.88%, 5.61%, 4.94% and 3.92%, respectively. In summary, BiSeNeXt achieves the highest performance across all evaluation metrics, except for leaf recall, where it ties with UNet++. Notably, it demonstrates a significant advantage in disease segmentation.
Table 9 summarizes the inference time, number of parameters and FLOPs of each segmentation model on the yam dataset. SegNeXt achieved the fastest inference time at 14.61 ms, followed by Segmenter at 16.16 ms and BiSeNeXt at 18.62 ms, while UNet++ was the slowest at 54.92 ms. BiSeNeXt has only 3.22M parameters, which is significantly fewer than the second-lightest model, DFL-UNet+CBAM, with 12.72M. In terms of FLOPs, BiSeNeXt requires just 20.82 G, whereas UNet++ reaches 552.70 G. Therefore, BiSeNeXt achieves the lowest parameter count and computational cost while maintaining a relatively fast inference speed, demonstrating the advantage in computational efficiency.
Figure 10 shows the mIoU performance of different methods during training. BiSeNeXt excels in training speed, accuracy and stability. It converges faster than other methods in the early stages and stabilizes above 93% mIoU in the later stages, outperforming other models. In addition, its curve shows minimal fluctuation, demonstrating the method’s efficiency and robustness.
Leaf occlusion poses a significant challenge to the accurate extraction of target leaf edge pixels. It hampers the model’s ability to clearly distinguish the boundaries between adjacent leaves, often leading to the misclassification of lesions on non-target leaves as target disease areas. Moreover, leaf curling further increases the difficulty of segmentation. Figure 11A shows the segmentation results of leaves and diseases under leaf occlusion conditions, and Figure 11B shows the corresponding manually annotated results. A comparison between Figure 11A and Figures 11C, D, F, H–J shows that all the evaluated models, including DeepLabV3+, DFL-UNet+CBAM, U-Net, PSPNet, SegNeXt and Swin Transformer, are affected by the presence of non-target leaves. This interference leads to the misidentification of lesion areas, where regions on non-target leaves are incorrectly extracted as diseased areas. In addition, as shown in Figures 11C, F, G, I–K, DeepLabV3+, U-Net, Segmenter, SegNeXt, Swin Transformer and Vision Transformer all show missing lesion areas in the segmentation results. Figure 11 demonstrates that, apart from the proposed method and UNet++, the remaining models generally suffer from varying levels of misclassification when dealing with occluded regions. Compared with other methods, as shown in Figures 11C–L, only the proposed method exhibits reliable robustness in accurately segmenting disease boundaries under leaf curling conditions.. Overall, the proposed method achieves superior accuracy in segmenting target leaves and extracting lesion areas.

Figure 11. Comparison of leaf and lesion segmentation methods under occlusion and leaf curling. (A) Original image. (B) Ground truth. (C) DeepLabV3+. (D) DFL-UNet+CBAM. (E) UNet++. (F) U-Net. (G) Segmenter. (H) PSPNet. (I) SegNeXt. (J) Swin Transformer. (K) Vision Transformer. (L) BiSeNeXt (Ours).

Figure 12. Comparison of different methods for leaf and dense lesion segmentation under uneven light conditions. (A) Original image. (B) Ground truth. (C) DeepLabV3+. (D) DFL-UNet+CBAM. (E) UNet++. (F) U-Net. (G) Segmenter. (H) PSPNet. (I) SegNeXt. (J) Swin Transformer. (K) Vision Transformer. (L) BiSeNeXt (Ours).
Figure 12 illustrates the impact of multiple segmentation models on leaf and dense disease segmentation under uneven lighting conditions. Figure 12A provides a typical schematic of leaf and disease segmentation in an uneven lighting environment, while Figure 12B demonstrates the ground truth for leaf and disease segmentation. As observed in Figures 12B–D, F, G, I–K, DeepLabV3+, DFL-UNet+CBAM, U-Net, Segmenter, SegNeXt, Swin Transformer and Vision Transformer struggle to accurately segment leaf edges under uneven illumination, leading to blurred contours and missing regions. As illustrated in Figures 12B, C, E, G), DeepLabV3+, UNet++ and Segmenter tend to misclassify leaves and disease regions in the presence of soil background interference. Figure 12 reveals that most models struggle to accurately segment densely distributed small lesions under uneven lighting conditions. In contrast, the method proposed in this study successfully identifies a larger portion of lesion areas, although some omissions remain. Additionally, it achieves high precision in segmenting leaf contours.
Figure 13 compares the performance of different methods on leaf and disease segmentation under the influence of raindrops and overlapping leaves. Figure 13A shows a typical example of dense leaf and disease segmentation under raindrop interference. It significantly challenges the model’s ability to distinguish pixels. Figure 13B presents the ground truth for leaf and disease segmentation. As shown in Figure 13G, Segmenter exhibits the poorest segmentation performance, failing to accurately delineate leaf and disease contours. Comparing Figures 13H–J shows that PSPNet, SegNeXt and Swin Transformer exhibit adhesion of neighboring lesions under raindrop interference. As shown in Figure 13, except for the proposed method, other models all fail to detect small lesions within the small yellow boxes under the influence of raindrops. As illustrated in Figure 13, except for the proposed method and UNet++, the other models miss small lesions within the yellow boxes under rain interference. Moreover, in scenarios where leaf overlap and raindrops coexist, these models are also prone to misclassifying background regions as leaves. Notably, as shown in Figure 13L, the proposed method demonstrates superior performance in segmenting leaf edges and small lesions, highlighting its enhanced robustness under challenging conditions.

Figure 13. Comparison of different methods on leaf and disease segmentation tasks under the influence of raindrops and overlapping leaves. (A) Original image. (B) Ground truth. (C) DeepLabV3+. (D) DFL-UNet+CBAM. (E) UNet++. (F) U-Net. (G) Segmenter. (H) PSPNet. (I) SegNeXt. (J) Swin Transformer. (K) Vision Transformer. (L) BiSeNeXt (Ours).
3.8 Generalization evaluation on apple leaf disease dataset
To verify the generalization ability of the proposed method on different crops, we evaluated it using an apple leaf disease image dataset. The dataset originates from Northwest A&F University and contains four common types of apple leaf disease spots: rust, gray spot, brown spot and alternaria. It contains 1222 images, including 334 of rust, 395 of gray spot, 215 of brown spot and 278 of alternaria. To ensure consistency in the experimental setup, all images were resized to 512×512 pixels, and the same data augmentation strategy was applied as in yam dataset. The dataset was randomly split into training, validation and test sets at a 7:2:1 ratio. As shown in Table 10, the BiSeNeXt model performs well on this dataset. Without fine-tuning, it achieved IoU scores of 98.90%, 96.98% and 80.66% in the background, leaf and spot regions, respectively. The mF1-Score reached 95.89%. Compared to the baseline model, BiSeNeXt improved the IoU in the spot region by 3.66%. This dataset differs significantly from the yam dataset in terms of crop type, leaf morphology, lesion distribution and shooting conditions. Therefore, our method demonstrates strong transferability. It has the potential to be generalized to multiple crop disease segmentation tasks.
4 Conclusions
In this paper, a yam disease segmentation dataset is constructed for the first time. It covers a wide range of environmental conditions, including indoor and outdoor settings, various lighting conditions and different weather. The dataset contains three common diseases: anthracnose, brown spot and gray spot. It provides a valuable data foundation and evaluation benchmark for yam disease research. In addition, this research proposes a yam leaf disease segmentation method based on the improved BiSeNetV2 network. The method addresses the challenges of low segmentation accuracy and high computational complexity in complex environments. To achieve this, three modules, DFEB, EAMA and PointRefine, are specifically designed to enhance segmentation performance. DFEB mitigates the information loss problem by employing dynamic sensory field convolution and pixel shuffling downsampling. This module enhances the recognition accuracy of leaf boundary pixels in complex scenes and reduces spot omissions. PointRefine adopts a point-to-point refinement strategy to restore fine details and edges. EAMA strengthens feature extraction in disease regions and alleviates spot adhesion via a multi-scale attention mechanism. The experimental results show that the proposed method reaches 97.04% IoU in leaf segmentation and 84.75% IoU in lesion segmentation. In contrast to existing methods, the proposed method enhances segmentation precision and effectively minimizes computational overhead. Additionally, cross-crop validation on an apple leaf disease dataset further demonstrates the model’s promising generalization ability without fine-tuning.
Despite the good results of this study, certain limitations remain. The current dataset does not yet include yam species and disease types from other regions. The model’s generalization ability on unseen samples remains to be validated. Segmentation accuracy requires improvement in complex scenarios such as extreme lighting conditions, dense foliage overlap and densely clustered small disease spots. In the future, we plan to expand the dataset to include environments such as fog and frost, and incorporate samples from other regions to improve the model’s generalization capabilities. Meanwhile, methods such as image enhancement and multi-scale feature extraction will also be integrated to further enhance the model’s segmentation accuracy in complex scenes. In addition, we plan to deploy the model on mobile platforms such as smartphones for real-time field monitoring of yam diseases. To address potential dynamic interferences such as device shaking, speed fluctuations, and changes in shooting distance, we will conduct field tests to evaluate their impact on segmentation performance and improve the model’s robustness. Moreover, mechanisms for speed monitoring, shake detection and distance alerts will be integrated to provide user feedback and reduce accuracy loss caused by improper operation.
Data availability statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://drive.google.com/drive/folders/1_ojcb_84TMbkZwYfm0dgsL1NjiGw7GRF?usp=sharing.
Author contributions
BL: Investigation, Data curation, Writing – review & editing, Conceptualization, Funding acquisition, Resources. YL: Data curation, Formal Analysis, Visualization, Validation, Methodology, Conceptualization, Writing – original draft, Writing – review & editing, Software, Investigation. DL: Data curation, Investigation, Writing – review & editing. JY: Writing – review & editing, Resources, Data curation.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the National Natural Science Foundation of China (Grant 42272178), the Key Scientific Research Project of Higher Education Institutions in Henan Province (Grant 24B520013), and the Special Project of Basic Scientific Research Operating Expenses (Natural Science) of Henan Polytechnic University (Grant NSFRF240508). The authors declare that financial support was received for the research, authorship and publication of this article.
Acknowledgments
We sincerely appreciate all the authors for their support and contributions to this manuscript.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Badrinarayanan, V., Kendall, A., and Cipolla, R. (2017). SegNet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39, 2481–2495. doi: 10.1109/tpami.2016.2644615
Baffour-Ata, F., Awugyi, M., Ofori, N. S., Hayfron, E. N., Amekudzi, C. E., Ghansah, A., et al. (2023). Determinants of yam farmers’ adaptation practices to climate variability in the ejura sekyedumase municipality, Ghana. Heliyon 9, e14090. doi: 10.2139/ssrn.4296408
Chen, L.-C. (2014). Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv preprint arXiv:1412.7062. doi: 10.1109/tpami.2017.2699184
Chen, L.-C. (2017). Rethinking atrous convolution for semantic image segmentation. arXiv preprint arXiv:1706.05587. doi: 10.1109/tpami.2017.2699184
Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K., and Yuille, A. L. (2017). DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 40, 834–848. doi: 10.48550/arXiv.1412.7062
Chen, Z., Shi, T., Zhang, X., Jia, K., Jiang, H., and Yuan, B. (2022). A hybrid leaf area index estimation method of dioscorea polystachya turczaninow using sentinel-2 vegetation indices. IEEE Trans. Geosci. Remote Sens. 60, 1–13. doi: 10.48550/arXiv.1706.05587
Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018). “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European conference on computer vision (ECCV), (Cham, Switzerland: Springer), vol. 801–818. doi: 10.1007/978-3-030-01234-2_49
Dai, W., Zhu, W., Zhou, G., Liu, G., Xu, J., Zhou, H., et al. (2024). Aisoa-ssformer: An effective image segmentation method for rice leaf disease based on the transformer architecture. Plant Phenomics 6, 218. doi: 10.34133/plantphenomics.0218
Diouf, M. B., Festus, R., Silva, G., Guyader, S., Umber, M., Seal, S., et al. (2022). Viruses of yams (dioscorea spp.): Current gaps in knowledge and future research directions to improve disease management. Viruses 14, 1884. doi: 10.3390/v14091884
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., et al. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929. doi: 10.48550/arXiv.2010.11929
Gao, J., Hu, X., Xiao, R., Luo, F., Tang, Y., Luo, J., et al. (2023a). The microbiome and typical pathogen multiplication, qualities changes of baoxing yam at different storage temperatures. LWT 188, 115402. doi: 10.1016/j.lwt.2023.115402
Gao, J., Zhao, W., Qin, Z., Chen, L., Li, M., Zhang, J., et al. (2023b). First report of root-knot nematode (meloidogyne incognita) on dioscorea opposita in Henan Province, China. Plant Dis. 107, 2556. doi: 10.1094/pdis-02-23-0359-pdn
Gogile, A., Kebede, M., Kidanemariam, D., and Abraham, A. (2024). Identification of yam mosaic virus as the main cause of yam mosaic diseases in Ethiopia. Heliyon 10, e26387. doi: 10.1016/j.heliyon.2024.e26387
Guo, M.-H., Lu, C.-Z., Hou, Q., Liu, Z., Cheng, M.-M., and Hu, S.-M. (2022). SegNeXt: Rethinking convolutional attention design for semantic segmentation. Adv. Neural Inf. Process. Syst. 35, 1140–1156. doi: 10.48550/arXiv.2209.08575
He, Y., Tong, C., Chen, H., Zhao, W., Zhan, L., Wang, R., et al. (2025). A rapid and visual detection method for alternaria alternata, the causal agent of leaf spot disease on yam, based on RPA-CRISPR/Cas12a. Physiol. Mol. Plant Pathol. 137, 102612. doi: 10.1016/j.pmpp.2025.102612
Hu, J., Shen, L., and Sun, G. (2018). “Squeeze-and-excitation networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Piscataway, NJ, United States: IEEE), 7132–7141. doi: 10.1109/CVPR.2018.00745
Kirillov, A., Wu, Y., He, K., and Girshick, R. (2020). “Pointrend: Image segmentation as rendering,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (Piscataway, NJ, United States: IEEE), 9799–9808. doi: 10.1109/cvpr42600.2020.00982
Kumar, A. and Sachar, S. (2023). Deep learning techniques in leaf image segmentation and leaf species classification: A survey. Wireless Pers. Commun. 133, 2379–2410. doi: 10.1007/s11277-02410873-2
Lee, H., Hwang, S. J., and Shin, J. (2019). Rethinking data augmentation: Self-supervision and selfdistillation. [Preprint]. doi: 10.48550/arXiv.1910.05872
Li, Y., Ji, S., Xu, T., Zhong, Y., Xu, M., Liu, Y., et al. (2023). Chinese yam (dioscorea): Nutritional value, beneficial effects, and food and pharmaceutical applications. Trends Food Sci. Technol. 134, 29–40. doi: 10.1016/j.tifs.2023.01.021
Li, X., Shao, Y., Hao, L., Kang, Q., Wang, X., Zhu, J., et al. (2024). The metabolic profiling of Chinese Yam fermented by saccharomyces boulardii and the biological activities of its ethanol extract in vitro. Food Sci. Hum. Wellness 13, 2718–2726. doi: 10.26599/fshw.2022.9250219
Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., et al. (2021). “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision (Piscataway, NJ, United States: IEEE), 10012–10022. doi: 10.1109/iccv48922.2021.00986
Liu, J., Zhao, B., and Tian, M. (2024). Asymmetric-convolution-guided multipath fusion for real-time semantic segmentation networks. Mathematics 12, 2759. doi: 10.3390/math12172759
Long, J., Shelhamer, E., and Darrell, T. (2015). “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Piscataway, NJ, United States: IEEE), 3431–3440. doi: 10.1109/CVPR.2015.7298965
Lu, X., Xiao, C., Zhou, K., Fu, L., Shen, D., and Dou, D. (2023). First report of nigrospora oryzae causing leaf spot on yam in China. Plant Dis. 107, 2256. doi: 10.1094/pdis-11-22-2545-pdn
Lu, L., Xiong, Q., Xu, B., and Chu, D. (2024). “Mixdehazenet: Mix structure block for image dehazing network,” in 2024 International Joint Conference on Neural Networks (IJCNN) (Piscataway, NJ, United States: IEEE), 1–10. doi: 10.1109/ijcnn60899.2024.10651326
Ouyang, D., He, S., Zhang, G., Luo, M., Guo, H., Zhan, J., et al. (2023). “Efficient multi-scale attention module with cross-spatial learning,” in ICASSP 2023–2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (IEEE), 1–5. doi: 10.1109/icassp49357.2023.10096516
Picon, A., San-Emeterio, M. G., Bereciartua-Perez, A., Klukas, C., Eggers, T., and NavarraMestre, R. (2022). Deep learning-based segmentation of multiple species of weeds and corn crop using synthetic and real image datasets. Comput. Electron. Agric. 194, 106719. doi: 10.1016/j.compag.2022.106719
Ronneberger, O., Fischer, P., and Brox, T. (2015). “U-Net: Convolutional networks for biomedical image segmentation,” in Medical image computing and computer-assisted intervention–MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, Part III, vol. 18. (Cham, Switzerland: Springer), 234–241. doi: 10.1007/978-3-319-24574-4_28
Shwetha, V., Bhagwat, A., Laxmi, V., and Shrivastava, S. (2024). A custom backbone UNet framework with DCGAN augmentation for efficient segmentation of leaf spot diseases in jasmine plant. J. Comput. Networks Commun. 2024, 5057538. doi: 10.1155/2024/5057538
Strudel, R., Garcia, R., Laptev, I., and Schmid, C. (2021). “Segmenter: Transformer for semantic segmentation,” in Proceedings of the IEEE/CVF international conference on computer vision (Piscataway, NJ, United States: IEEE), 7262–7272. doi: 10.48550/arXiv.2105.05633
Tariq, H., Xiao, C., Wang, L., Ge, H., Wang, G., Shen, D., et al. (2024). Current status of yam diseases and advances of their control strategies. Agronomy 14, 1575. doi: 10.3390/agronomy14071575
Wang, T., Wang, Y., Geng, Z., Wei, J., Chang, Y., Zhu, M., et al. (2024). Identification and pathogenicity of colletotrichum truncatum causing yam anthracnose–a new record in China. Physiol. Mol. Plant Pathol. 131, 102246. doi: 10.1016/j.pmpp.2024.102246
Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., and Hu, Q. (2020). “Eca-net: Efficient channel attention for deep convolutional neural networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (Piscataway, NJ, United States: IEEE), 11534–11542. doi: 10.1109/cvpr42600.2020.01155
Wang, B., Yang, M., Cao, P., and Liu, Y. (2025). A novel embedded cross framework for high-resolution salient object detection. Appl. Intell. 55, 277. doi: 10.1007/s10489-024-06073-x
Woo, S., Park, J., Lee, J., and Kweon, I. (2018). “CBAM: Convolutional block attention module,” in Proceedings of the proceedings of the European conference on computer vision (ECCV). (Cham, Switzerland: Springer). doi: 10.1007/978-3-030-01234-2_1
Yang, Z., Sun, L., Liu, Z., Deng, J., Zhang, L., Huang, H., et al. (2025). Fatdnet: A fusion adversarial network for tomato leaf disease segmentation under complex backgrounds. Comput. Electron. Agric. 234, 110270. doi: 10.1016/j.compag.2025.110270
Yu, C., Gao, C., Wang, J., Yu, G., Shen, C., and Sang, N. (2021). BiSeNet v2: Bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vision 129, 3051–3068. doi: 10.1007/s11263-021-01515-2
Zhang, Y., Huang, S., Zhou, G., Hu, Y., and Li, L. (2023c). Identification of tomato leaf diseases based on multi-channel automatic orientation recurrent attention network. Comput. Electron. Agric. 205, 107605. doi: 10.1016/j.compag.2022.107605
Zhang, X., Li, D., Liu, X., Sun, T., Lin, X., and Ren, Z. (2023a). Research of segmentation recognition of small disease spots on apple leaves based on hybrid loss function and cbam. Front. Plant Sci. 14. doi: 10.3389/fpls.2023.1175027
Zhang, X., Liu, C., Yang, D., Song, T., Ye, Y., Li, K., et al. (2023b). RFAConv: Innovating spatial attention and standard convolutional operation. arXiv preprint arXiv:2304.03198. doi: 10.48550/arXiv.2304.03198
Zhang, S. and Zhang, C. (2023). Modified U-Net for plant diseased leaf image segmentation. Comput. Electron. Agric. 204, 107511. doi: 10.1016/j.compag.2022.107511
Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017). “Pyramid scene parsing network,” in Proceedings of the proceedings of the IEEE conference on computer vision and pattern recognition. (Piscataway, NJ, United States: IEEE). doi: 10.1109/cvpr.2017.660
Zhao, W., Zhao, H., Wang, H., and He, Y. (2022). Research progress on the relationship between leaf senescence and quality, yield and stress resistance in horticultural plants. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.1044500
Zhou, Z., Rahman Siddiquee, M. M., Tajbakhsh, N., and Liang, J. (2018). “Unet++: A nested u-net architecture for medical image segmentation,” in Deep learning in medical image analysis and multimodal learning for clinical decision support: 4th international workshop, DLMIA 2018, and 8th international workshop, ML-CDS 2018, held in conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, proceedings, vol. 4. (Cham, Switzerland: Springer), 3–11. doi: 10.1007/978-3-030-00889-5_1
Zhou, Y., Zhou, H., and Chen, Y. (2024). An automated phenotyping method for chinese cymbidium seedlings based on 3D point cloud. Plant Methods 20, 151. doi: 10.1186/s13007-024-01277-1
Keywords: yam leaf segmentation, disease spot segmentation, DFEB, EAMA, PointRefine
Citation: Lu B, Lu Y, Liang D and Yang J (2025) BiSeNeXt: a yam leaf and disease segmentation method based on an improved BiSeNetV2 in complex scenes. Front. Plant Sci. 16:1602102. doi: 10.3389/fpls.2025.1602102
Received: 28 March 2025; Accepted: 15 July 2025;
Published: 05 August 2025.
Edited by:
Thomas Thomidis, International Hellenic University, GreeceReviewed by:
Nitin Goyal, Central University of Haryana, IndiaYan Guo, Henan Academy of Agricultural Sciences, China
Copyright © 2025 Lu, Lu, Liang and Yang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Bibo Lu, bHViaWJvQGhwdS5lZHUuY24=