ORIGINAL RESEARCH article

Front. Oncol., 13 March 2025

Sec. Cancer Imaging and Image-directed Interventions

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1522399

3D segmentation of uterine fibroids based on deep supervision and an attention gate

  • 1. Center of Radiation Oncology, Ganzhou Cancer Hospital, Ganzhou, Jiangxi, China

  • 2. State Key Laboratory of Ultrasound in Medicine and Engineering, College of Biomedical Engineering, Chongqing Medical University, Chongqing, China

  • 3. Department of Radiology, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China

Article metrics

View details

6

Citations

1,5k

Views

509

Downloads

Abstract

Introduction:

The segmentation of uterine fibroids is very important for the treatment of patients. However, uterine fibroids are small and have low contrast with surrounding tissue, making this task very challenging. To solve these problems, this paper proposes a 3D DA- VNet automatic segmentation method based on deep supervision and attention gate.

Methods:

This method can accurately segment uterine fibroids in MRI images by convolutional information. We used 3DVnet as the underlying network structure and added a deep monitoring mechanism in the hidden layer. We introduce attention gates during the upsampling process to enhance focus on areas of interest. The network structure is composed of VNet, deep supervision module and attention gate module. The dataset contained 245 cases of uterine fibroids and was divided into a training set, a validation set, and a test set in a ratio of 6:2:2. A total of 147 patients' T2-weighted magnetic resonance (T2WI) images were used for training, 49 for validation, and 49 patients' MR Images were used for algorithm testing.

Results:

Experimental results show that the proposed method achieves satisfactory segmentation results. Dice similarity coefficient (DSC), intersection ratio (IOU), sensitivity, precision and Hausdorff distance (HD) were 0.878, 0.784, 0.879, 0.885 and 11.180 mm, respectively.

Discussion:

This shows that our proposed method can improve the automatic segmentation accuracy of magnetic resonance image (MRI) data of uterine fibroids to a certain extent

1 Introduction

Uterine fibroids are one of the most common benign tumors in women, with an incidence of 20-40% in women during their reproductive period (1). Uterine fibroids can lead to serious morbidity, such as heavy menstrual bleeding and pelvic pressure (2). In addition, they pose a serious threat to women’s health and affect women’s quality of life. Traditional treatment for fibroids is hysterectomy, which can cause physical and emotional pain in women. In recent years, high-intensity focused ultrasound (HIFU) has been widely used and successfully in the treatment of uterine fibroids due to its noninvasive characteristics (3–5). However, whether a fibroids patient is treated with traditional surgery or HIFU, fibroids magnetic resonance image (MRI) segmentation is required before surgery, as MRI is currently considered to be the most accurate imaging technique for detecting and locating fibroids (6). T2-weighted imaging (T2WI) is very important for the diagnosis of gynecological diseases, and examined images can clearly show diseased tissue of the uterus, such as uterine fibroids and evidence of cervical cancer and endometrial cancer. T1-weighted images (T1WIs) enable better observation of anatomical structures. Segmentation of uterine fibroids is usually performed on T2WI MRI images. This needs to be done manually by one or two experienced physicians; manual segmentation of uterine fibroids takes a lot of time, and the segmentation results obtained by doctors vary from person to person, so uterine fibroid MRI segmentation is still a challenging task. The reasons are as follows: (1) The contrast between uterine fibroids and other surrounding tissue is low and the boundary is difficult to distinguish. (2) The area of uterine fibroids on MRI is small, with little available information. (3) The positions of the uterine fibroids are not fixed and it is difficult to segment them.

In recent years, deep learning has become very popular. An increasing number of researchers have begun to use deep learning to classify, recognize and segment medical images, including performing segmentation of uterine fibroids. Regarding deep neural networks, Kurata et al. (7) proposed an improved UNet to segment MRI images of uterine fibroids, replacing the ReLU in each layer in the original UNet with leaky ReLU. A dropout layer was added, batch_size was set to 15, 8 layers of downsampling were used, and finally the Dice similarity coefficient of all uterine fibroids was determined to be 0.820. Tang et al. (8) proposed AR-UNet, which used the ResNet101 deep neural network as the front end of feature extraction to extract semantic information from the image, reducing the number of layers and improving the precision of segmentation. They also introduced an attention gate module between up-sampling and down-sampling and incorporated UNet to build a network structure. Finally, the Dice similarity coefficient of MRI segmentation of uterine fibroids reached 0.904. Zhang et al. (9) proposed HIFUnet, a network with a ResNet101 backbone, global convolutional network (GCN) module, deep multiple atrous convolutions (DMAC) module, sampling, a cascading layer and an output layer. It was used for accurate segmentation of the uterus, uterine fibroids and spine on MRI, and the Dice similarity coefficient of segmentation of uterine fibroids was 0.835. However, these methods have common shortcomings. First, they cannot directly process 3D images, and second, the segmentation accuracy is not high. The existing methods usually convert 3D data into 2D slices through dimensionality reduction or image slicing before training the deep learning model, thus losing considerable image spatial information, which is not conducive to the effective segmentation of 3D images, so the segmentation accuracy is not high. Maike et al (10)proposed a method of automatic segmentation of uterus and uterine fibroids on MRI with 3D nn Unet, and achieved good segmentation results. Therefore, to obtain more in-depth information and better meet the needs of clinical diagnosis, it is necessary to propose a more effective 3D image processing method.

The purpose of this study was to explore the feasibility of automatic 3D VNet segmentation of uterine fibroids based on deep supervision and an attention gate. We applied a 3D deep supervision attention VNet (3D DA-VNet) and optimized some hyperparameters to achieve automatic segmentation. The contributions of this paper are as follows.

  • Compared to existing segmentation methods, we propose a 3D segmentation method to segment uterine fibroids.

  • To solve the problem of disappearance of the gradient and slow convergence, a deep supervision mechanism is introduced.

  • Deep supervision and an attention gate are integrated to improve segmentation accuracy.

The structure of this paper is as follows. Section 2 details the proposed materials and methods, including data preprocessing and model description. Section 3 is mainly about the results and discussion, including parameter settings and evaluation indicators. The conclusion is given in Section 4.

2 Materials and methods

2.1 Dataset

Data from January 2013 to December 2018 in the first affiliated hospital of the Haifu Minimally Invasive and Non-invasive Treatment Center of the Chongqing Medical University for the ablation treatment of 245 patients with uterine fibroids. The image data parameters are shown in Table 1.

Table 1

parametersT2WI
TR(ms)3740
TE(ms)106.6
FOV(cm)98.1 38
Matrix512 512
Slices22
Slice thickness(mm)6
Slice gap(mm)8
Nex2

Imaging parameters of MRI T2WI.

TR, Repetition time; TE, Echo time; FOV, Field of view; Nex, Number of excitations.

2.2 Experimental parameters

The experiment was carried out on a TensorFlow 8G NVIDIA GeForce RTX 3070 graphics card, CUDA version 11.3.1. Furthermore, the initial learning rate was set at 2e-4, and the epoch was set at 600. As the epoch increased, the learning rate decreased exponentially at a rate of 0.999. Due to large 3D data and the limitation of device memory, the batch size could only be set to 1; the loss function was Dice (Dice Loss) (11), Adam (12) was used as optimizer, the momentum was 0.9, and PReLU (13) was used as activation function.

2.3 Preprocessing

The labels were manually delineated by a physician with 3 years of experience using ITK-SNAP on axial T2WI. The obtained regions of interest were used as the gold standard for segmentation of uterine fibroids. Image preprocessing included normalization, resampling, filling, cropping and random noise. The purpose of normalization was to make the grey values of each image in the training set have the same distribution. Resampling was performed to normalize the voxels of different sizes in the image to the same size. The actual spatial size represented by a single voxel in different images was inconsistent. Because the convolutional neural network only operates in the voxel space, it ignores the size information in the actual physical space. To address this difference, it was necessary to resize different image data in the voxel space to ensure that the actual physical space represented by each voxel was consistent across different image data. The image was then filled and cut to an input size of 128x128x48.

2.4 Deep surpervision module

Deep supervision is also called relay supervision. As shown in Figure 1, auxiliary loss functions are added to some intermediate hidden layers of the deep neural network as network branches to supervise the trunk network. Moreover, additional loss functions are added to the middle part of the network, and loss functions at different positions are summed by coefficients. The purpose is to train the features more fully and solve the problems of gradient disappearance and slow convergence of the deep neural network. As a training strategy, deep supervision was proposed in 2014 through deep supervision nets (DSNs) (14), which can improve the directness and transparency of the hidden layer learning process. To reduce the adverse effects of unstable gradient changes, we propose using display supervision to train the hidden layers in 3D DA-VNet. Specifically, we first use additional deconvolution steps to amplify some low-level and middle-level features. We then use softmax functions on these full-size characteristic volumes and obtain improved predictions. For these branch prediction results, we calculate their errors from the manual segmentation results. These auxiliary losses are integrated with the losses of the last layer to stimulate the backpropagation of gradients for more efficient parameter updates with each iteration.

Figure 1

We connect the volume features directly through the path to the last output layer as the primary network. Let be the weight of layer L of the primary network. Using W=(, …, ), priority is given to the network weight set, and P(| ;w) represents the probability prediction of voxel after the softmax function in the last output layer. The negative log-likelihood loss function is expressed as:

where represents the training data, represents the target label of , and belongs to . Additionally, we create an auxiliary intensive prediction layer called the branch network. Deep supervision is introduced through the branch network. When deep supervision is introduced in the m-th hidden layer, is used to represent the weight of the first m layers in the main network, and the weight of intensive prediction can be connected to the characteristic volume of the m-th layer through . Then, the auxiliary loss function of deep supervision can be expressed as follows:

Finally, we learn the weight and all using the backpropagation algorithm by minimizing the following objective function:

where represents the balance weight of , which decays during learning. is the division of all the hidden layers with a deep supervision function. The first term corresponds to the prediction of the output in the last output layer, and the second term represents deep supervision. In each training iteration, the network’s input is the large-capacity data, and simultaneously, the error is backpropagated from these different weight losses.

The effectiveness of deep supervision can be proven through the following considerations. First, Qi et al. (15), in a liver segmentation challenge and a heart and large vessel segmentation challenge, used a deep supervision network to achieve a higher speed than the most advanced methods of competitive segmentation, thus proving the effectiveness of the deep supervision network. Zeng et al. (16) proposed a 3D UNet fully convolutional network with deep supervision for the segmentation of the proximal femur in 3D magnetic resonance (MR) images, and the experimental results proved the effectiveness of deep supervision. Zhu et al. (17), Yang et al. (18) and Bo et al. (19) applied deep supervision modules in different network structures to automatically segment the prostate gland. The final results proved that the deep supervision module could effectively address the optimization problem of gradient disappearance or explosion when training the 3D model, accelerate the convergence speed and improve the recognition ability.

2.5 Attention gate module

The attention gate (20) is shown in Figure 2, where represents the feature map of the upward sampling and represents the feature map of the downwards sampling of the same layer. The two feature maps are adjusted to the same size and added together. The attention coefficient can be obtained through ReLU, 1×1×1 convolution, a sigmoid function and resampling, and then through multiplication by , can be obtained. The attention gate module can better ensure that attention is given to the prominent area and suppress the irrelevant background area, and it can be effectively embedded in the VNet network.

Figure 2

2.6 Network framework

VNet (21) usually consists of an input layer, a convolution layer, a lower sampling layer, an upper sampling layer and an output layer. The convolutional layer in the structure is used to learn image features, and local connections and weight sharing are used to reduce the number of parameters and the computational complexity. With the deepening of the network, layer-by-layer convolution can extract more abstract image features. As one of the key steps of pattern recognition, the quality of feature extraction directly affects the accuracy of image recognition. Through the study of VNet, we propose the 3D DA-VNet structure, which is more suitable for the 3D segmentation of MRI data of uterine fibroids. The overall framework of 3D DA-VNet is shown in Figure 3. In the upsampling stage, we combine the upsampling features with the downsampling features to introduce an attention gate. The attention gate module can better ensure that attention is given to the prominent area and suppress the irrelevant background area, and it can be effectively embedded in the VNet network. At the same time, in the upsampling stage, a deep supervision module is introduced to continuously upsample the output of each layer until the outputs are the same size as the training image. Then, a softmax function can be chosen to generate the contour probability maps followed by labelling each voxel. Accordingly, the losses of these intermediate layers together with that of the final output layer are combined for gradient backpropagation, which is used to identify more effective parameters with each iteration.

Figure 3

The parameter details of 3D DA-VNet are provided in Table 2. In the input layer, we input an image with a size of 128 × 128 × 48 and then convolve it with a 3 x 3 x 3 convolution kernel and activate it with PReLU. On this basis, standardization and a dropout rate of 0.01 are applied. The subsequent convolution operation is the same. Then, through downsampling, the stride of the convolution kernel is set to (1,2,2,2,1), and the deep features of the images are extracted. After continuous convolution and downsampling, the fifth layer is obtained, and its size is 8 × 8 × 3. After the deconvolution operation, the stride of the deconvolution kernel is (1,2,2,2,1), and the node is processed using the PReLU activation function. Therefore, the structure of the processing model retains the compatibility of the original network and lays the foundation for rapid iterative optimization. In the upsampling stage, the concatenation calculation is performed after each deconvolution operation. With this operation, the last layer uses softmax to predict values 0 and 1 for each pixel. Finally, the size of the image is recovered at 128 × 128 × 48.

Table 2

LayerInput SizeComponents
Left-S 1128*128*483*3*3,16, stride (1, 2, 2, 2, 1)
Left-S 264*64*243*3*3,32, stride (1, 2, 2, 2, 1)
Left-S 332*32*123*3*3,64, stride (1, 2, 2, 2, 1)
Left-S 416*16*63*3*3,128, stride (1, 2, 2, 2, 1)
Left-S 58*8*33*3*3,256, stride (1, 2, 2, 2, 1)
Right-S 416*16*63*3*3,128, stride (1, 2, 2, 2, 1)
Right-S 332*32*123*3*3,64, stride (1, 2, 2, 2, 1)
Right-S 264*64*243*3*3,32, stride (1, 2, 2, 2, 1)
Right-S 1128*128*483*3*3,16, stride (1, 2, 2, 2, 1)
output128*128*483*3*3,16, stride (1, 1, 1, 1, 1)

Parameter details of the proposed 3D DA-VNet.

2.7 Evaluation metrics

We divided the data of 245 cases of uterine fibroids into a training set, testing set, and validation set at a ratio of 6:2:2, among which 147 cases were used for training, 49 cases were used for testing, and 49 cases were used for validation. Training and testing were carried out at the same time. After the model training was completed, it was verified by the validation set. The output of 3D DA-Vet was the prediction result of uterine fibroids. We used a single case as the input data to evaluate the performance of the model.

Before high-intensity focused ultrasound treatment of uterine fibroids, it is necessary to accurately locate uterine fibroids, so the accuracy of segmentation directly influences the therapeutic effect. However, for a pelvic image, the fibroid region only accounts for a very small part of the image, which often causes the segmented part to be neglected by the network, the output of the network being biased towards the background, and learning falling into a local extremum; ultimately, accurate results cannot be obtained. To avoid this problem, Spatial overlap was quantified using the Dice Similarity Coefficient (DSC) (22, 23). However, it’s important to note that DSC lacks a clear definition in cases where both compared volumes contain zero positive voxels, resulting in division by zero. The DSC calculation is as follows:

The IOU score is a standard performance measure for segmentation problems. Given an image, the IOU score gives the similarity between the predicted regions and ground-truth regions presented in the image, and it is defined by the following equation:

Sensitivity (24) measures the positive part of the voxel in the real background, which measures the ability to segment the fibroid region of the uterus:

Precision (25), also known as positive predictive value, refers to the accuracy of the segmentation of uterine fibroids:

where TP, FP and FN are the probabilities of true positives, false positives and false negatives, respectively. These evaluation indicators are all based on area and are sensitive to the divided internal filling area, while the Hausdorff distance (HD) (26) is based on distance and is sensitive to the divided boundary. The Hausdorff distance is a measurement used to describe the degree of similarity between two sets of points. Here, we use it to evaluate the relationship between the segmentation results and the real fibroid boundary distance.

where gt represents the ground truth, seg represents the segmentation result, A and B represent the voxels of the ground truth and the segmentation result, respectively, and D(a, b) represents the Euclidean distance between A and B.

3 Results and discussion

3.1 Effect of different convolution kernels and model structure

The convolution kernel is one of the most important concepts in deep learning. It has the advantage of weight sharing and translational invariance, and Yao Jin et al. (27) proposed that its size affects the number of model parameters and the information of extracted images. Therefore, we designed 3 × 3 × 3 and 5 × 5 × 5 convolution kernels. In addition, minor adjustments to the overall structure will affect the final result, including the number of convolutions for each layer. We evaluated our proposed 3D DA-VNet on datasets using DSC, IOU, sensitivity, precision, and HD, and present the results of the 6 experiments in Table 3.

Table 3

the Model StructureConvolution Kernel SizeDSCIOUsensitivityprecisionHD(mm)
(1,2,3,3,3,3,3,2,1)3*3*30.876 0.0380.781 0.0590.867 0.0700.8930.07011.34 7.216
5*5*50.858 0.0400.754 0.0600.858 0.0740.870 0.08212.01 6.842
(1,2,2,2,2,2,2,2,1)3*3*30.870 0.0390.772 0.0600.863 0.0760.888 0.07512.07 7.650
5*5*50.862 0.0530.773 0.0660.875 0.0770.875 0.07511.45 7.163
(1,2,3,3,2,2,2,2,2)3*3*30.8780.0370.7840.0570.8790.0720.885 0.06711.187.181
5*5*50.857 0.0370.752 0.0570.863 0.0850.865 0.08012.36 7.840

Quantitative evaluation results of the dataset.

Bold numbers indicate the best results.

3.2 The influence of different network training on the results

To validate the proposed method, we tested T2WI MRI of 49 patients and segmented different shapes of uterine fibroids. To compare the segmentation results with the ground truth, we calculated DSC, IOU, sensitivity, accuracy, and HD, and Table 4 shows the results of comparing the algorithm proposed in this paper with 2D UNet, 2D VNet, 3D UNet, 3D VNet, 3D deep supervision VNet (3D DS-VNet) and 3D attention VNet. The DSC of our proposed method reached 0.878. The IOU was 0.784, the sensitivity was 0.879, the precision was 0.885, and the HD was 11.18 mm. Due to the large amount of 3D MRI segmentation data, the average segmentation time of a uterine fibroid was 7 seconds. Figure 4 compares the segments that result in the segmentation. As seen from Figure 4, compared to 3D DS-VNet and 3D attention VNet, 3D DA-VNet has a better segmentation effect at the fibroids’ boundary and has a more accurate segmentation of uterine fibroids, which gives it advantages in the image processing of uterine fibroids. For fibroids with irregular shapes and small sizes, the segmentation effect is better.

Table 4

MethodDSCIOUSensitivityPrecisionHD(mm)
2D Unet0.725 0.1620.614 0.1330.724 0.1410.789 0.15328.38 12.760
3D Unet0.740 0.0890.626 0.1150.739 0.0800.805 0.14327.90 10.496
2D Vnet0.783 0.1400.697 0.1290.787 0.1480.857 0.16320.85 11.014
3D Vnet0.833 0.0620.740 0.0740.848 0.0840.866 0.10613.70 7.164
3D DS-Vnet0.846 0.0510.750 0.0700.863 0.0760.862 0.09612.55 7.218
3D attention Vnet0.862 0.0430.772 0.0580.873 0.0730.878 0.07512.23 7.767
3D DA-Vnet0.8780.0370.7840.0570.8790.0720.8850.06711.187.181

Testing results of uterine fibroids segmentation on different networks.

The best results (expressed as mean ± standard deviation) are shown in bold.

Figure 4

3.3 Discussion

We compared the size and number of different convolution kernels, as shown in Table 3. From the final results, we chose the size of the convolution kernel as 3x3x3, with (1, 2, 3, 3, 2, 2, 2, 2, 2) as the structure. We also compared different 2D and 3D networks, as well as UNet and VNet networks. The final results in Table 4 show that the segmentation effect of our proposed 3D DA-VNet is the best, and they also prove that the deep supervision module and the attention gate module are effective. Additionally, Figure 4 confirms that our predicted target regions have a good overlap with the standard manual segmentation results given. To better understand the results of our segmentation, we compared the results of the segmentation of uterine fibroid MRI data with those of other studies. The corresponding comparisons are shown in Table 5. These are traditional segmentation methods, including the fuzzy C-means algorithm, split-and-merge, level set segmentation, etc. Their segmentation results are also good, but these traditional methods require complex preprocessing and postprocessing of the data. Segmentation is not performed synchronously, which makes the segmentation process inefficient. In conclusion, although these conventional methods have some merits in terms of performance, they show some practical limits in a clinical setting. Table 6 gives a comparison with some deep learning segmentation methods. The existing deep learning methods for segmenting uterine fibroids all segment 2D slice data, which loses considerable image spatial information and is not conducive to the effective segmentation of 3D images, so the segmentation accuracy is not high. To obtain more in-depth information, to better meet the needs of clinical diagnosis, and to propose a more effective 3D image processing method, we propose the automatic segmentation of uterine fibroids using 3D DA-VNet based on deep supervision and attention gates. The deep supervision module allows features to be more fully trained by deep feature supervision, thus improving the segmentation performance. The attention gate module improves segmentation performance by ensuring that more attention is given to the region of interest while suppressing irrelevant background regions, and the final DSC value in the testing set is 0.878.

Table 5

AuthorsYearDataset sizeDataset sourceAlgorithmEvaluation metrics
N. Ben-Zadok et al (28)20094privateLevel set segmentation,
User feedback
TPE=0.999
A. Fallahi et al (29)20105privatefuzzy C-means algorithm and
morphological operations,modified possibilistic fuzzy c-mean
SI=0.799
A. Fallahi et al (30)201110privateFuzzy C-Mean (FCM) method and morphological operationsSI=0.800
H. Khotanlou et al (31)201415privateChan–Vese level set method
geometric shape prior model
SI=0.877
Xu M et al (32)201542privatesplit-and-mergeSI=0.876
Militello C et al (33)201515privateUnsupervised Fuzzy C-Means clustering and iterative optimal threshold selection algorithmsSI=0.887
Rundo L et al (34)201614privateDataset normalization
Region splitting and merging
Seed-region refinements
SI=0.876
Rundo L et al (35)201918privateIterative Optimal Threshold Selection (IOTS)
Split-and-Merge(SM),Region Growing(RG)
DSC=0.873
DSC=0.875
Ning G et al (36)2020320privateCNN, encoder and decoder pathsDSC=0.812

Comparison of traditional methods for segmenting uterine fibroids.

Table 6

AuthorsYearDataset sizeDataset sourceNetworkDSCIOUPrecisionRecall (Sensitivity)
Ours2022249private3D DA-Vnet0.8780.7840.8850.879
Kurata Y et al (7)2019122privateU-net0.820———
Tang et al (8)202093privateAR-Unet0.9040.844—0.886
Zhang et al (9)2020297privateHIFUNet0.835—0.8450.837
Wang et al (37)2024550privatenn-Unet0.956———

Comparison of methods for deep learning segmentation of uterine fibroids.

However, we can see from Table 6 that compared with AR-Unet and nn-Unet results, our results are not outstanding. The main reason for our analysis lies in the difference of images. We used transverse magnetic resonance images, while Tang (8) and Wang (37) used sagittal magnetic resonance images. In anatomy, sagittal images can more clearly show the long axis of the uterus and the uterine lumen-fibroid interface, while the transverse section is easy to truncate the fibroid shape, and the boundary of the transverse expanded fibroids is blurred, making segmentation difficult. Our study mainly focuses on the efficacy of HIFU ablation in the treatment of uterine fibroids. We will use different MRI sequences, which will all have transverse images, but not necessarily sagittal images. Therefore, we choose to develop a model for segmenting transverse images, so as to better combine clinical treatment in our subsequent studies. On the other hand, the lack of data is also a factor that makes the results low.

Judging from the results, the algorithm we proposed demonstrates a certain efficacy in the segmentation of uterine fibroids. However, our study is not without limitations that may impact the generalizability and accuracy of the findings. Firstly, the sample size of our data is relatively small; specifically, 245 MRI images may be insufficient for training deep learning models given the complexity inherent in MRI data. The ability to generalize this model to larger and more diverse datasets remains uncertain. Secondly, our magnetic resonance imaging (MRI) dataset is limited to a single series; this study relies solely on T2-weighted imaging. Uterine fibroids may exhibit different characteristics across other MRI sequences or imaging modalities, which constrains the clinical applicability of this model when applied under varying imaging conditions. Finally, this study does not compare with current state-of-the-art segmentation techniques and may underestimate the performance advantages or limitations of these techniques. While the methods employed in this study perform well under certain conditions, more room for improvement and potential challenges may be found compared to the latest techniques. Therefore, future research will incorporate these into the research scope to ensure the advanced nature and optimization potential of the method. Additionally, due to the presence of significant surrounding tissue around uterine fibroids, caution must be exercised when utilizing high-intensity focused ultrasound for ablation procedures so as not to damage adjacent structures. Therefore, multi-organ segmentation encompassing uterine fibroids along with related anatomical structures such as the uterus and spine represents a more aligned approach with actual clinical needs—this also constitutes a promising direction for future research endeavors.

4 Conclusions

This study uses deep learning to achieve automatic segmentation of uterine fibroids magnetic resonance data. Most of the previously proposed deep learning models first convert 3D MRI data into 2D image slices and then use them to train and optimize the network, which wastes considerable time and does not accord with the 3D characteristics of the images themselves. For the 3D DA-VNet proposed in this paper, the input layer is a 3D matrix, which can speed up training progress and improve training efficiency. In addition, this paper introduces a deep supervision module and attention gate. The use of the deep supervision module allows the layers to be more fully trained and addresses the depth gradient disappearance and the slow convergence speed of the neural network, and integration of the attention gate can better ensure focus on the area of interest and suppression of irrelevant background regions. The experimental results for the same data set and platform show that the accuracy of the proposed method is significantly improved compared to other segmentation methods. It is believed that with improvement in segmentation accuracy, automatic segmentation of uterine fibroids can help doctors make more accurate judgments, improve work efficiency, and reduce the rate of misdiagnosis.

Statements

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

ZL: Writing – original draft, Writing – review & editing. CS: Conceptualization, Data curation, Software, Validation, Writing – original draft. CL: Data curation, Formal analysis, Investigation, Project administration, Writing – original draft. FL: Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    MavrelosDBen-NagiJHollandTHooWNaftalinJJurkovicD. The natural history of fibroids. Ultrasound Obstet Gynecol. (2010) 35:238–42. doi: 10.1002/uog.v35:2

  • 2

    StewartEA. Uterine fibroids. Lancet. (2010) 357:293–8. doi: 10.1016/S0140-6736(00)03622-9

  • 3

    KennedyJE. High-intensity focused ultrasound in the treatment of solid tumours. Nat Rev Cancer. (2005) 5:321–7. doi: 10.1038/nrc1591

  • 4

    RenXLZhouXDZhangJHeGBHanZHZhengMJet al. Extracorporeal ablation of uterine fibroids with high-intensity focused ultrasound: imaging and histopathologic evaluation. J Ultrasound Med. (2007) 26:201–12. doi: 10.7863/jum.2007.26.2.201

  • 5

    WangWWangYWangTWangJWangLTangJ. Safety and efficacy of US-guided high-intensity focused ultrasound for treatment of submucosal fibroids. Eur Radiol. (2012) 22:2553–8. doi: 10.1007/s00330-012-2517-z

  • 6

    ZhaoWPChenJYZhangLLiQQinJPengSet al. Feasibility of ultrasound-guided high intensity focused ultrasound ablating uterine fibroids with hyperintense on T2-weighted MR imaging. Eur J Radiol. (2013) 82(1):e43–e49. doi: 10.1016/j.ejrad.2012.08.020

  • 7

    KurataYNishioMKidoAFujimotoKYakamiMIsodaHet al. Automatic segmentation of the uterus on MRI using a convolutional neural network. Comput Biol Med. (2019) 114:103438. doi: 10.1016/j.compbiomed.2019.103438

  • 8

    Chun-mingTDongLXiangY. MRI image segmentation system of uterine fibroids based on AR-unet network. Am Sci Res J Engineering Technol Sci. (2020) 71:1–10.

  • 9

    ZhangCShuHYangGLiFWenYZhangQet al. HIFUNet: multi-class segmentation of uterine regions from MR images using global convolutional networks for HIFU surgery planning. IEEE Trans Med Imaging. (2020) 39:3309–20. doi: 10.1109/TMI.42

  • 10

    TheisMTongucTSavchenkoONowakSBlockWReckerFet al. Deep learning enables automated MRI-based estimation of uterine volume also in patients with uterine fibroids undergoing high-intensity focused ultrasound therapy. Insights Imaging. (2023) 14:1. doi: 10.1186/s13244-022-01342-0

  • 11

    LiXSunXMengYLiangJWuFLiJ. Dice loss for data-imbalanced NLP tasks. arXiv: Comput Lang. (2019) 465–76. doi: 10.18653/v1/2020.acl-main.45

  • 12

    KingmaDPBaJ. Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980. (2014). doi: 10.48550/arXiv.1412.6980

  • 13

    HeKZhangXyuRenSSunJ. Delving deep into rectifiers: surpassing human-level performance on imageNet classification. International Conference on Computer Vision(ICCV). (2015), 1026–34. doi: 10.1109/ICCV.2015.123

  • 14

    LeeC-YXieSGallagherPZhangZTuZ. Deeply-supervised nets. Proceedings of the 2015 IEEE International Conference on Artificial Intelligence and Machine Learning (ICML), (2015), pp. 1011–9. doi: 10.48550/arXiv.1409.5185

  • 15

    DouQiYuLChenHJinYYangXQinJet al. 3D deeply supervised network for automated segmentation of volumetric medical images. Med Image Anal. (2017) 41:40–54. doi: 10.1016/j.media.2017.05.001

  • 16

    ZengGZhengG. Deep learning-based automatic segmentation of the proximal femur from MR images. Adv Exp Med Biol. (2018) 1093:73–9. doi: 10.1007/978-981-13-1396-7_6

  • 17

    ZhuQDuBTurkbeyBPChoykeLYanP. (2017). Deeply-supervised CNN for prostate segmentation, in: 2017 International Joint Conference on Neural Networks (IJCNN), (New York, USA: IEEE). pp. 178–84.

  • 18

    LeiYTianSHeXWangTWangBPatelPet al. Ultrasound prostate segmentation based on multidirectional deeply supervised V-Net. Med Phys. (2019) 46:3194–206. doi: 10.1002/mp.2019.46.issue-7

  • 19

    WangBLeiYTianSWangTLiuYPatelPet al. Deeply supervised 3D fully convolutional networks with group dilated convolution for automatic MRI prostate segmentation. Med Phys. (2019) 46:1707–18. doi: 10.1002/mp.2019.46.issue-4

  • 20

    OktayOSchlemperJoFolgocLLLeeM JHeinrichMMisawaKet al. Attention U-net: learning where to look for the pancreas. arXiv:1804.03999. (2018) 429–37. doi: 10.48550/arXiv.1804.03999

  • 21

    MilletariFNavabNAhmadiSA. (2016). V-net: fully convolutional neural networks for volumetric medical image segmentation, in: Proceedings of 2016 Fourth International Conference on 3d Vision (3dv), (New York, USA: IEEE). pp. 565–71.

  • 22

    DiceLR. Measures of the amount of ecologic association between species. Ecology. (1945) 26:297–302. doi: 10.2307/1932409

  • 23

    SudreCHLiWVercauterenTOurselinSJorge CardosoM. Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations. Deep Learn Med Image Anal Multimodal Learn Clin Decis Suppor. (2017) 2017:240–8. doi: 10.1007/978-3-319-67558-9_28

  • 24

    TrevethanR. Sensitivity, specificity, and predictive values: foundations, pliabilities, and pitfalls in research and practice. Front Public Health. 5:307. doi: 10.3389/fpubh.2017.00307

  • 25

    MonteiroFCCampilhoAC. Performance evaluation of image segmentation Vol. 5. ICIAR. Springer, Berlin, Heidelberg, (2006) p. 248–59.

  • 26

    HenriksonJT. Completeness and total boundedness of the Hausdorff metric. MURJ. (1999) 1:69–80.

  • 27

    JinYYangGFangYLiRXuXLiuYet al. 3D PBV-Net: An automated prostate MRI data segmentation method. Comput Biol Med. (2021) 128:104160. doi: 10.1016/j.compbiomed.2020.104160

  • 28

    Ben-ZadokNRiklin-RavivTKiryatiN. (2009). Interactive level set segmentation for image-guided therapy, in: 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, (Boston, MA, USA). pp. 1079–82.

  • 29

    FallahiAPooyanMKhotanlouHHashemiHFirouzniaKMA. (2010). Uterine fibroid segmentation on multiplan MRI using FCM, MPFCM and morphological operations, in: 2010 2nd International Conference on Computer Engineering and Technology, China. pp. V7–1-V7-5.

  • 30

    FallahiAPooyanMGhanaatiHOghabianMAKhotanlouHShakibaMet al. Uterine segmentation and volume measurement in uterine fibroid patients’ MRI using fuzzy C-mean algorithm and morphological operations. Iran J Radiol. (2011) 8:150–6. doi: 10.5812/kmp.iranjradiol.17351065.3142

  • 31

    KhotanlouHFallahiAOghabianMAPooyanM. Segmentation of uterine fibroid on mr images based on chan–vese level set method and shape prior model. Biomed Eng: Appl Basis Commun. (2014) 26:1450030. doi: 10.4015/S1016237214500306

  • 32

    XuMZhangDYangYLiuYYuanZQinQ. A split-and-merge-based uterine fibroid ultrasound image segmentation method in HIFU therapy. PloS One. (2015) 10:e0125738. doi: 10.1371/journal.pone.0125738

  • 33

    MilitelloCVitabileSRundoLRussoGMidiriMGilardiMC. A fully automatic 2D segmentation method for uterine fibroid in MRgFUS treatment evaluation. Comput Biol Med. (2015) 62:277–92. doi: 10.1016/j.compbiomed.2015.04.030

  • 34

    RundoLMilitelloCVitabileSCasarinoCRussoGMidiriMet al. Combining split-and-merge and multi-seed region growing algorithms for uterine fibroid segmentation in MRgFUS treatments. Med Biol Eng Comput. (2016) 54:1071–84. doi: 10.1007/s11517-015-1404-6

  • 35

    RundoLMilitelloCTangherloniARussoGLagallaRMauriGet al. Computer-assisted approaches for uterine fibroid segmentation in MRgFUS treatments: quantitative evaluation and clinical feasibility analysis. Quantifying Process Biomed Behav Signals. (2019) 38(12):2687–98. doi: 10.1007/978-3-319-95095-2_22

  • 36

    NingGZhangXZhangQWangZLiaoH. Real-time and multimodality image-guided intelligent HIFU therapy for uterine fibroid. Theranostics. (2020) 10:4676–93. doi: 10.7150/thno.42830

  • 37

    WangTWenYWangZ. nnU-Net based segmentation and 3D reconstruction of uterine fibroids with MRI images for HIFU surgery planning. BMC Med Imaging. (2024) 24:233. doi: 10.1186/s12880-024-01385-3

Summary

Keywords

uterine fibroid, MRI segmentation, deep supervision, attention gate, deep learning

Citation

Liu Z, Sun C, Li C and Lv F (2025) 3D segmentation of uterine fibroids based on deep supervision and an attention gate. Front. Oncol. 15:1522399. doi: 10.3389/fonc.2025.1522399

Received

23 December 2024

Accepted

24 February 2025

Published

13 March 2025

Volume

15 - 2025

Edited by

Lia Morra, Polytechnic University of Turin, Italy

Reviewed by

Mehmet Dalmis, Meta, Netherlands

Palash Ghosal, Sikkim Manipal University, India

Updates

Copyright

*Correspondence: FaJin Lv,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics