Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol., 24 July 2025

Sec. Cancer Imaging and Image-directed Interventions

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1608963

Risk assessment of thyroid nodules with a multi-instance convolutional neural network

Da YuDa Yu1Tingting SongTingting Song2Yancheng YuYancheng Yu1Hebin ZhangHebin Zhang1Feng GaoFeng Gao1Zirong Wang*Zirong Wang3*Jiacheng WangJiacheng Wang4
  • 1Department of Ultrasound, The Affiliated Hospital of Hangzhou Normal University, Hangzhou, Zhejiang, China
  • 2Department of Radiology, The Fourth People’s Hospital of Harbin, Harbin, Heilongjiang, China
  • 3Department of Radiology, The Affiliated Hospital of Hangzhou Normal University, Hangzhou, Zhejiang, China
  • 4Department of Computer Science, School of Informatics, Xiamen University, Xiamen, China

Objectives: Ultrasonography is the primary imaging modality for evaluating thyroid nodules, and artificial intelligence (AI) has advanced the automated diagnosis of thyroid cancer. However, existing AI-assisted methods often suffer from limited diagnostic performance.

Methods: In this study, we propose a novel multi-instance learning (MIL) convolutional neural network (CNN) model tailored for ultrasound-based thyroid cancer diagnosis. The model extracts nodule-level ultrasound features from instance-level images using CNNs, and employs an attention mechanism to assign importance scores and aggregate features across instances. This enables effective feature extraction and localization of key instance features, facilitating risk assessment of thyroid nodules. The dataset consists of ultrasound images from 2000 patients at the Affiliated Hospital of Hangzhou Normal University, collected between 2018 and 2024. The images were divided into training (75%, 1500 patients) and testing (25%, 500 patients) sets. The model's performance was evaluated using metrics, including accuracy, precision, recall, F1-Score, and AUC. To assess the statistical significance of the model’s performance relative to other methods, a paired t-test was conducted based on the prediction results.

Results: The performance of the model developed in this study was evaluated and compared with popular ultrasound image classification models for thyroid nodules. The model outperformed the other two classification models (accuracy 0.8386±0.0334, 0.7999±0.0188, 0.7839±0.0267; precision 35 0.8512±0.0301, 0.9039±0.0154, 0.9267±0.0235; recall0.8427±0.0313, 0.7497±0.0163, 0.6987±0.0249; F1-Score 0.8380±0.0344, 0.8196±0.0178, 0.7967±0.0251; AUC 0.8900±0.0309, 0.8851±0.0124, 0.6340±0.0200), where values are under 95% confidence interval. Statistical analysis showed that the performance differences were statistically significant (p <0.0001).

Conclusions: These results demonstrate the effectiveness and clinical utility of the proposed MIL-CNN framework in non-invasively stratifying thyroid nodule risk, supporting more informed clinical decisions and potentially reducing unnecessary biopsies and surgeries. Codes will be available at GitHub - rrrr-ops/Thyroid-AI.

1 Introduction

In 2020, the World Health Organization’s International Agency for Research on Cancer (GLOBOCAN) Global Cancer Statistics showed that thyroid cancer had a worldwide incidence of approximately 586,000 cases, which ranked 9th in incidence, with the highest percentage being papillary thyroid cancer, which accounted for approximately 84% of all thyroid cancers (1, 2). Thyroid cancer may occur in any sex, with women accounting for approximately 75% of all thyroid cancer patients (3). Chinese scholars counted the epidemiological characteristics of 9,662 cases of thyroid cancer from 2014 to 2019 and found that the number of thyroid cancer cases showed an increasing trend year by year, and the growth rate of 2017–2019 was significantly higher than that of the previous three years (4). In 2019, 230,000 new cases and 45,000 deaths were associated with thyroid cancer worldwide, with China accounting for 16.71% and 15.88% of cases, respectively. China has the highest incidence and mortality rates of thyroid cancer worldwide, and the incidence of thyroid cancer in Zhejiang Province has risen sharply (57). In parallel with economic development and advancements in early detection, thyroid nodules have drawn increasing clinical attention (8). Thyroid cancer has a significant impact on public health, and despite its relatively low mortality rate, the rising incidence of thyroid cancer has increased the demand for healthcare resources and put tremendous pressure on the public health system. With the aging of the population, the incidence of thyroid cancer is likely to increase further, posing an even greater challenge to the public health system.

The Expert Committee of the Superficial Organs and Vascular Group of the Division of Ultrasound Medicine of the Chinese Medical Association (CMA) formulated the Chinese Guidelines for Ultrasound Malignancy Risk Stratification of Thyroid Nodules (C-TIRADS) in 2020, which is more in line with China’s actual situation (9). This classification system integrates multiple ultrasound features of nodules, such as nodule morphology, margins, internal echogenicity, and calcification, to classify nodules into different categories, from category 1 (no nodule, 0% risk of malignancy) to category 6 (biopsy-confirmed malignancy), with an increased risk of malignancy. The C-TIRADS classification provides critical guidance for clinical decision-making. For nodules classified as category 2 by C-TIRADS (0% malignancy probability), fine-needle aspiration (FNA) biopsy is usually not necessary, whereas for nodules in categories 4B to 5 (10% to >90% malignancy probability), FNA is required for further evaluation (10). The C-TIRADS classification helps clinicians determine which patients are suitable for surgical treatment and which patients provide more conservative treatment options with active monitoring for follow-up or ultrasound-guided thermal ablation (11). Accurate nodule stratification with C-TIRADS facilitates early identification of high-risk nodules and timely therapeutic intervention, thereby improving patient outcomes. In particular, low-risk papillary thyroid microcarcinoma (PTMC) may be managed with active surveillance to avoid overtreatment, whereas high-risk nodules require prompt surgical intervention (12).

C-TIRADS plays an important role in the diagnosis and treatment of thyroid cancer; however, its practical application faces some challenges. First, sample heterogeneity was a key issue. Thyroid nodules exhibit diverse characteristics, including cystic nodules, homogeneous hypoechoic nodules, nodular goiter, calcified nodules, microcalcifications, and diffuse microcalcifications, which can be benign, malignant, or suspected malignant, with varying degrees of severity (13). Since the ultrasound manifestations of thyroid nodules may be similar, and their biological characteristics and treatment options may be very different, C-TIRADS should be able to accurately differentiate between different types of thyroid nodules. Second, the subjectivity of the diagnosis is challenging. The diagnosis of ultrasound images is highly subjective, and there are differences in the interpretation of ultrasound features of thyroid nodules by sonographers with different experiences and inter- and intra-observer variations in the assessment of ultrasound metrics of the nodules, which may affect the accuracy and consistency of the C-TIRADS classification (14, 15). These subjective variations underscore the clinical need for automated diagnostic systems that can provide standardized and reproducible assessments across institutions and operators. In addition, the C-TIRADS provides detailed management recommendations and clinical guidance; however, many thyroid surgeries in China are based on ultrasound diagnostic reports rather than FNA results, sometimes in combination with other relevant clinical bases. Appropriate adjustments are required for practical applications according to regional differences, national characteristics, cultural traditions, and medical systems (9).

Artificial intelligence (AI) has been widely applied to medical tasks including detection, classification, and segmentation (16, 17). In particular, with the rapid advancement of AI technologies, machine learning and deep learning have become promising approaches for the automatic classification of thyroid nodules, and are now extensively used in ultrasound imaging, fine-needle aspiration (FNA) biopsy, and thyroid surgery (18). The general steps of AI for the diagnosis of thyroid nodules include: (i) modeling optimized preprocessing of ultrasound images and selecting the region of interest from them, (ii) applying feature extraction methods for training to classify thyroid nodules, and (iii) diagnosis of benign or malignant nodules. AI is increasingly used in the field of ultrasound with the continuous progress of algorithm iterations and accumulation of clinical data (1921). Compared to traditional thyroid cancer diagnostic methods, such as FNA biopsy, AI promises to provide a non-invasive alternative for classifying thyroid cancers with higher accuracy by analyzing their histopathological features, clinical manifestations, and prognosis. For example, Li et al. collected more than 40,000 patients with thyroid nodules and jointly used two models commonly used in deep learning, ResNet and DarkNet, to differentiate between benign and malignant thyroid nodules, and obtained an AUC of more than 0.9 (22).These findings suggest that AI has the potential for cross-center use in the diagnosis of thyroid nodules. Peng et al. developed an AI deep learning model (ThyNet) to differentiate between thyroid cancer and benign thyroid nodules (23), and the results showed that the AUC of the features obtained by the ThyNet model-assisted diagnosis was 0.922, which is significantly higher than that of the physician’s 0.839. The percentage of patients with fine-needle aspiration biopsies decreased from 61.9% to 35.2% and the percentage of missed thyroid cancer lesions decreased from 18.9% to 17.0% when using the ThyNet-assisted system. The ThyNet model showed better diagnostic specificity and accuracy than specialized physicians.

However, traditional AI-assisted thyroid cancer diagnostic techniques have some limitations that need to be addressed. First, the diagnostic technique based on a single ultrasound image is limited by the one-sidedness of the information, which is easily affected by a variety of factors, such as imaging quality, patient position, and physician experience, leading to misdiagnosis or missed diagnosis (2325). Second, traditional techniques are difficult to apply to highly heterogeneous ultrasound image features presented by different patients or even the same patient in different disease courses, making it difficult to cope with complex and changing conditions. In addition, annotation based on a single image not only increases the burden on medical resources but also easily leads to inconsistent diagnostic results due to physician qualifications and regional differences (26). Therefore, these models face challenges in clinical applications because they lack the ability to integrate information regarding the patient as a whole.

In recent years, multiple-instance learning (MIL) techniques have provided new ideas for solving these problems (2729). MIL is a subfield of machine learning, a paradigm that learns data relationships from a set of instances and bags, and by virtue of the unique structure of bag-instances, is able to not only solve the problem of computational resources in large-scale images. Existing MIL methods have made significant progress in the field of pathology images, achieving substantial improvements in computational efficiency, instance feature extraction, and instance aggregation mechanisms, making them an important for the pathological image analysis. However, the application of MIL in ultrasonography remains to be explored.

Unlike conventional deep learning methods based on single-frame ultrasound images, our model uses a multiple-instance learning framework to aggregate information across frames, enhancing robustness and aligning more closely with clinical practice. The model can receive a collection of patient ultrasound examination video frames as input, and by synthesizing and analyzing the information in these image frames, it can comprehensively evaluate the patient’s condition and provide accurate thyroid cancer diagnosis results. In addition, we compared the feature extraction capabilities of multiple classification networks in the ultrasound diagnosis of thyroid cancer to further improve diagnostic accuracy and model robustness. We further analyzed the interpretation of the model predictions and compared the computational costs, which has proven that our model has potential to be applied in practice.

2 Materials and methods

2.1 Thyroid nodules risk prediction model based on multi- instance learning

2.1.1 Multi-instance learning framework and problem definition

MIL, as a weakly supervised learning method, is particularly suitable for dealing with the situation where there are multiple relevant instances in the image data but only the overall label is known. Traditional single-instance learning assumes that each training sample has an explicit label, whereas multi-instance learning assumes that each training sample consists of multiple instances, and only the overall label is known. MIL requires that the data be represented in a package-instance structure; when there are positive instances within the package, the whole package is labeled as a positive package; otherwise, the package and all its instances are negative.

The rationale for selecting MIL in this study, as opposed to other algorithms, is driven by the objective of performing thyroid nodule risk assessment based on the patient’s comprehensive information. Specifically, the ultrasound examination process of a patient can be collected as a set of images, referred to as a bag, where each image is considered an instance. The number of instances per patient can vary flexibly according to the actual examination conditions, without the need for consistency. The attention mechanism in the MIL framework is capable of handling dynamically changing instance numbers, effectively adapting to this variable-length input structure. The risk reflected in a single image may not always correspond to the patient’s actual risk level. For instance, certain images from high-risk patients may not directly indicate any evident risk. Given the multifaceted nature of thyroid nodule risk assessment, relying solely on individual images may lead to inaccurate predictions. In this context, MIL is employed to analyze all images of a patient collectively, synthesizing the information from each instance to yield a more accurate and holistic risk assessment (30, 31). Therefore, the architecture of MIL makes it relevant to the context of medical image analysis and risk assessment.

In the MIL model, the instance feature extraction network and instance feature aggregation mechanism are two core components. Instance feature extraction networks are typically designed based on CNN and transformers. In this study, we adopt a CNN-based backbone network. Compared to transformer-based models, CNNs offer two key advantages: (1) they are more effective at capturing local spatial features through multi-layer convolution and pooling operations, which is particularly important for texture-rich ultrasound images (32); and (2) their relatively simple architecture and fewer parameters help reduce the risk of overfitting—a critical consideration given the size of our dataset (2,000 patients, including 1,500 in the training set). The learning objective of these networks is to extract the representative features from each instance. By contrast, instance feature aggregation mechanisms are used to integrate features from multiple instances into package-level features. Commonly used aggregation methods include global average pooling (GAP), maximum pooling (MP), and attention mechanisms. The attention mechanism plays an important role in this study as it can guide the model to focus on key image segments with diagnostic significance, thus improving the accuracy of diagnosis and interpretability of the model. In addition, the feature fusion strategy is optimized to enhance the accuracy of bag-level prediction using methods such as multi-layer perceptron (MLP) to effectively fuse instance features.

In this paper, we propose a thyroid nodules risk prediction algorithm based on multi-instance learning, which aims to reduce reliance on image-level labels and focuses on the extraction and analysis of key instance features. In this algorithm, the set of ultrasound images of each patient was considered a package, whereas each image in the image set was treated as an instance, and the overall risk label of the package was determined based on the comprehensive examination results without the need to label each instance individually. For thyroid nodules risk prediction, the model can extract discriminative key representations from ultrasound images using an improved instance feature extraction network based on convolutional neural networks. Subsequently, in the instance feature aggregation mechanism, the model can select the key ultrasound images that match the clinical experience as the basis for judgment and locate the location and extent of the lesions in the key images. During the training process, the MIL model derives bag-level labels by learning the instance-level information, thereby effectively addressing the challenges of data heterogeneity and complexity. Combined with the MIL approach, the model can simultaneously analyze multiple related images of ultrasound species, jointly analyze the patient’s condition, and provide a more accurate diagnosis.

2.1.2 Model architecture

In this study, we constructed a package containing the ultrasound images of each patient’s thyroid ultrasound results. The model designs a VGG13 based instance feature extraction module and implemented the aggregation of instance features using an attention mechanism.

First, in the example feature extraction module, the model employs the VGG13 network pretrained on ImageNet, a large-scale natural image dataset, and is further trained on thyroid ultrasound images. VGG13 is known for its unique convolutional layer stacking structure, which progressively extracts features at different levels in an image through multiple consecutive convolutional and pooling layers. It relies on successive 3×3 convolutional kernels and rectified linear unit activation functions to deepen the network. Each convolutional layer is typically followed by a pooling layer to reduce the spatial resolution of a feature map. To enhance the adaptability of the VGG13 network to thyroid ultrasound images, we introduced various data enhancement techniques, such as random rotation, scaling, and panning, during the training process to improve the generalization ability of the model and enhance its robustness to data noise. The instance feature extraction module extracts the deepest feature maps of VGG13, and after pooling them, the features of each instance are obtained for subsequent attentional learning. After obtaining multiple instance features, it is crucial to know how to aggregate these instance features to obtain patient-level thyroid cancer diagnosis features. To this end, the algorithm refers to the classical attention-based multi-instance learning method AB-MIL (33) which learns the importance of each instance in the final diagnosis through matrix operations and tanh(.) function and assigns different weights to each instance to obtain patient-level feature representations through weighted summation.

Specifically, given a set of instance features H=h1,h2,,hnN×512, where N is the number of instances and each hi is a feature vector with 512 dimensions, the attention weights are computed using two fully connected layers followed by a softmax function:

A=SoftmaxW2·TanhW1HT+b1+b2

where W1512×128 and W2128×1 are learnable weight matrices, and b1128 and b2128 are bias terms. The attention weights of each instance can be simplified as:

ai=exp(W2T·Tanh(W1hi+b1)+b2))j=1Nexp(W2T·Tanh(W1hj+b1)+b2)

This approach automatically assigns higher weights to important instances, thereby improving overall discriminative power. The weighted feature vector M is then obtained by:

M=i=1Naihi

Eventually, the aggregated feature vectors are passed through the fully connected layer and dichotomized (benign or malignant) using the sigmoid activation function.

2.2 Experiment

2.2.1 Datasets

The thyroid ultrasound image data used in this study were obtained from the Affiliated Hospital of Hangzhou Normal University, and the images were acquired using ultrasound equipment (Mindray DC-8 and Mindray Resona 7EXP) with different numbers of ultrasound images obtained in a single examination. A total of 2000 patients were enrolled between January 2018 and July 2024. Ultrasound reports were collected for each examination. According to the C-TIRADS diagnostic criteria (9), the dataset includes 1000 patients with category 3 nodules, 975 patients with category 4 nodules, and 25 patients with category 5 nodules. Nodules classified as C-TIRADS category 3 were considered low-risk, while those in categories 4 and 5 were considered high-risk. A total of 16229 ultrasound images from 2000 patients with thyroid nodules were collected for this study. Of these, 12,283 images from 1500 patients were used for the training set, 750 patients had low-risk nodules and 750 patients had high-risk nodules. The test set consisted of 3946 ultrasound images from 500 patients, of which 250 patients were low-risk nodules and 250 patients were high-risk nodules. The mean age of the patients in the training and test sets was (49.83 ± 11.72) and (46.23 ± 11.46) years, respectively. The demographic characteristics of the two data sets are shown in Table 1.

Table 1
www.frontiersin.org

Table 1. Demographic data for two data sets.

In this study, all images from a single examination shared the same label, indicating that the examination had a high-risk nodule. All examinations were classified into two categories: high-risk and low-risk for thyroid nodules and were labeled according to the ultrasound report. The labeling process was performed by a single sonographer. The study was conducted in accordance with the Helsinki Declaration, and approved by the Ethics Committee of our institution (Approval number: 2024 (E2) -KS-160).

2.2.2 Data preprocessing

The dataset was randomly divided into a training set and a test set at a ratio of 3:1, where 75% of the data were used for training the model and 25% of the data were used for the final evaluation of the model. The entire dataset division process ensured that different patient categories were distributed in both training and test sets. All the images were preprocessed uniformly before being fed into the model to ensure stability and robustness. First, the images were resized to 512 × 512 pixels. To enhance the generalization ability of the model, the images were enhanced using random flipping and contrast enhancement. These transformations simulate different imaging conditions and increase the robustness of the model. Finally, the images were normalized to the pixel values.

2.2.3 Training strategies

The model was trained using the Adam optimizer with an initial learning rate of 0.00005 and a batch size of 1 (one patient per batch). The loss function was chosen as the Binary Cross-Entropy (BCE). NVIDIA RTX3090 was used for model training. The entire training process was conducted in an end-to-end manner, taking approximately 12 hours to complete over 100 epochs.

3 Results

3.1 Assessment of indicators

Several commonly used classification evaluation indices were used in this study to comprehensively assess the performance of the model for thyroid risk diagnosis. The accuracy rate was used to measure the proportion of overall correct classifications of the model, and the precision indicated the proportion of actual positive cases when the model predicted a positive case, thereby reflecting the model’s ability to determine a positive class. On the other hand, recall measures the proportion of all actual positive cases correctly predicted by the model, reflecting the model’s ability to detect high-risk patients. The F1-Score serves as the reconciled mean of precision and recall, which combines the precision and recall of the model. In addition, we plotted receiver operating characteristic (ROC) curves to evaluate the model’s ability to discriminate between dichotomous classification tasks under different thresholds. The higher area under the curve (AUC) value, the better is the overall classification ability of the model. Together, these evaluation metrics measure the performance of the model in thyroid nodules risk diagnosis and validate its accuracy, reliability, and ability to detect high-risk cases from various perspectives. Additionally, we utilized Cohen’s d to quantify the effect size and assess the magnitude of differences between the proposed model and comparison methods. Cohen’s d is a measure of the standardized difference between two means. It provides insight into the practical significance of the model’s improvements. In this study, a Cohen’s d value around 0.2 indicates a small effect size, around 0.5 represents a medium effect size, and around 0.8 corresponds to a large effect size. This effect size measurement helps to evaluate not only statistical significance but also the practical impact of the improvements made by our model.

3.2 Analysis of feature extraction networks

In this experiment, we used several different convolutional neural networks as feature extractors and comparatively evaluated the performance of these networks in the risk diagnosis of thyroid nodule ultrasound images under an attention-based multi-instance learning framework. The selected networks include the ResNet family, AlexNet, the VGGNet family, the MobileNet family, and the DenseNet family, which represent typical current deep learning models of different complexity and structural design.

As can be observed from the results in Table 2, VGG13 performed well on several evaluation metrics and became the best feature extraction network for this task. Specifically, VGG13 significantly outperforms the other networks in accuracy (0.8386), precision (0.8512), recall (0.8427), and F1-Score (0.8380), demonstrating the advantages of its deep convolutional structure in capturing the features of ultrasound images of the thyroid nodule. Notably, the comparison between VGG13 and other models, such as ResNet34, AlexNet, VGG11, MobileNetv3-small, and DenseNet161, yielded statistically significant differences (p<0.0001). The performance of VGG16 is slightly inferior to that of VGG13, even though its F1 Score is closer in performance, but it is slightly inferior in precision and recall. This may be due to the fact that VGG16 has more convolutional layers, resulting in a model that is more prone to overfitting. In contrast, ResNet50, another deep residual network, performed well in terms of accuracy (0.7375), precision (0.7375), and recall (0.6875). The residual structure of ResNet helps avoid the problem of gradient vanishing in deep networks, making it a more stable performer for processing complex image data. However, the results also indicate that ResNet50’s performance is statistically inferior to VGG13 (p<0.0001). On the contrary, AlexNet performs poorly on all metrics, and its F1-Score is only 0.3623, which is much lower than that of the other models. This indicates that AlexNet’s shallow structure cannot effectively extract deep features in thyroid ultrasound images, resulting in its weak classification ability. The large effect size (Cohen’s d = 0.3903 compared to VGG13) underscores the substantial gap between these two models. The MobileNet series and DenseNet series perform in the middle of the list, especially MobileNet as a lightweight network, although its precision (0.6420) and recall (0.6416) are not outstanding, it has the limited computational resources still has some advantages. The DenseNet series performed well in terms of recall (0.6560) through the densely connected design of the feature graph, but the overall performance failed to outperform VGG13 and ResNet50. Although VGG13 demonstrates clear numerical advantages over ResNet18, VGG16, DenseNet169, and DenseNet201 across various metrics, it does not exhibit statistically significant differences. Furthermore, the Cohen’s d effect size is small (less than 0.02), indicating that, despite numerical differences, these variations may have minimal impact on patient predictions. However, due to the numerical superiority of VGG13, we still consider it to be a more suitable feature extraction network compared to the others.

Table 2
www.frontiersin.org

Table 2. Results of feature extraction networks (VGG13 as the reference model, α=0.05 for statistical significance).

As shown in Figure 1, the AUC-ROC curves further demonstrate the performance of each model under different classification thresholds. We measured the comprehensive classification ability of the models using the AUC value, and the results showed that VGG13 had the highest AUC value of 0.8900, which further proved the excellent performance of VGG13 in thyroid nodules ultrasound image feature extraction.

Figure 1
Five ROC curve plots comparing network architectures. ResNets show clear separation, with ResNet18 having an AUC of 0.83. AlexNet has the lowest performance with an AUC of 0.37. VGGNets perform well, VGG13 achieves an AUC of 0.89. MobileNets show moderate performance, MobileNetV2 scoring an AUC of 0.68. DenseNets show similar performance across variations, with AUCs around 0.68-0.69. Diagonal reference lines indicate random performance.

Figure 1. ROC curves of compared methods on the test set.

3.3 Comparative with SOTA networks

To further validate the effectiveness of the proposed model, we designed an experiment to compare its performance with other models commonly used for thyroid nodules ultrasound image classification. The selected comparison models included LoGoNet (22) and ThyNet (23), both which are currently popular thyroid nodules classification models. All the models were trained on the same dataset and evaluated using the same metrics.

As can be seen from the experimental results in Table 3, LoGoNet has accuracy (0.7839) and precision (0.9267), but its low recall (0.6497) suggests that it may miss some malignant cases. The combination of high precision and low recall may reflect the model’s superior performance in avoiding misdiagnosis of benign cases but at the cost of overlooking high-risk patients. In contrast, ThyNet has a better recall rate than LoGoNet, but its precision rate is slightly lower, suggesting that ThyNet performs better in capturing malignant cases but also has a higher false positive rate. Overall, ThyNet strikes a balance between recall and precision but is still inferior to our proposed model.

Table 3
www.frontiersin.org

Table 3. Results of our model and comparison methods. (VGG13 as the reference model, α=0.05 for statistical significance).

Our proposed multi-instance learning model based on VGG13 performs well on several metrics, especially with significant advantages in recall (0.8427) and AUC (0.8900). This indicates that the model is more reliable for identifying high-risk thyroid nodules patients and effectively reduces the leakage rate. Recall, which measures the proportion of all actual positive cases that are correctly predicted as positive by the model, is particularly important for thyroid nodules classification. A higher recall means that the model can capture more malignant cases, thus reducing the risk of underdiagnosis. The experimental results show that the recalls of LoGoNet, ThyNet, and our model were 0.6497, 0.7497, and 0.8427, respectively. Our proposed model significantly outperformed the other two comparative models in terms of recall, indicating that it is more reliable in detecting high-risk patients and reduces the likelihood of missed diagnosis. Thus, our model has important implications for clinical diagnosis and is particularly superior in avoiding missed diagnoses, which may pose serious health risks.

However, a high recall usually leads to a decrease in precision because more positive class predictions may contain false positives. Therefore, whether the model has achieved a balance between precision and recall is a key question that we need to analyze further.The F1-Score is the reconciled mean of precision and recall, and combines the measures of how well the model is balanced on both metrics.The F1-Score is particularly important in unbalanced classification tasks, such as the thyroid cancer classification task, as it captures the model’s performance in identifying positive cases and avoiding false positives as a trade-off. The experimental results show that the F1-Score of LoGoNet, ThyNet, and our model are 0.7967, 0.8196, and 0.8380, respectively.Our proposed model also leads in F1-Score, indicating that the model achieves a better balance between precision and recall, which avoids a large number of false positives and is able to efficiently capture malignant cases.

In addition to the above metrics, statistical significance and effect size further validate the superiority of the proposed model. The p-values for both LoGoNet and ThyNet compared to our model are less than 0.0001, indicating that the differences are statistically significant. This highlights the reliability of our model’s improvements in classification performance. Cohen’s d, which measures the effect size, provides additional insights into the practical significance of these improvements. For LoGoNet and ThyNet, the Cohen’s d values are 0.1824 and 0.3762, respectively, suggesting small to medium effect sizes. This indicates that the improvements achieved by our model over ThyNet are more pronounced than those over LoGoNet.

Moreover, compared to SOTA networks, our model diagnoses at the patient level rather than on individual images, avoiding misdiagnoses due to poor image quality. Our attention mechanism operates across different images from the same patient, selecting critical diagnostic images and filtering out low-quality or misleading ones. These design choices enhance diagnostic reliability and improve model interpretability by highlighting clinically meaningful image segments.

In summary, the model proposed in this study demonstrates a comprehensive performance superior to that of existing models in the thyroid nodules ultrasound image classification task, especially in the identification and classification accuracy of high-risk patients, showing its great potential in clinical applications.

3.4 Analysis of attention mechanism

To evaluate the effectiveness of the attention mechanism in our MODEL, we conducted an ablation study comparing different instance aggregation strategies: max pooling, mean pooling, and our proposed attention-based aggregation. As shown in Table 4, the attention mechanism significantly outperformed both baseline pooling methods across all evaluation metrics. Specifically, our model achieved the highest accuracy.

Table 4
www.frontiersin.org

Table 4. Ablation studies of attention mechanism.

These results demonstrate that replacing simple pooling operations with an attention mechanism allows the model to adaptively assign different importance weights to each instance, thereby capturing more informative and discriminative features for patient-level classification. The superior performance confirms the critical role of attention in our MIL-based architecture.

3.5 Interpretable analysis

To increase the interpretability of the model, we used Grad-CAM (39) to visualize and analyze the decision-making process of the model. Grad-CAM is a visualization technique that highlights the regions of an input image that most influence the model’s prediction, by leveraging the gradient information flowing into the final convolutional layer. Using Grad-CAM, a heat map of attention for each ultrasound image can be generated, as shown in Figure 2, thus revealing the key areas that the model focuses on when making classification decisions.

Figure 2
Three pairs of images display ultrasound scans and corresponding thermal images. The top row shows grayscale ultrasound scans of tissues, each with different formations. The bottom row illustrates the same areas with color-coded thermal imaging, highlighting temperature variations with red indicating warmer areas and blue cooler regions.

Figure 2. The heat map of attention for each ultrasound image using Grad-CAM. (A): A includes ultrasonogram and heat map, ultrasonogram shows a mixed echogenic thyroid nodule, in which the solid portion has irregular morphology, unclear demarcation from the surrounding parenchyma, and multiple internal microcalcifications, diagnosed as a C-TIRADS category 4 nodule, heat map shows areas of dense aggregation with warm colors; (B): B includes ultrasonogram and heat map, ultrasonogram shows an isoechoic thyroid nodule with irregular morphology, an aspect ratio of greater than 1, and posterior echogenic attenuation, diagnosed as a C-TIRADS category 4 nodule, heat map shows warm-toned color as an area of dense aggregation; (C): C includes ultrasonogram and heat map, ultrasonogram shows a hypoechoic nodule with regular morphology and homogeneous internal echogenicity, diagnose as a C-TIRADS category 3 nodule, and heat map shows warm-toned color as an area.

Based on the thyroid nodule ultrasound image features, the warm-toned region in the visualized image is the most important part that can be identified by Multi-instance learning convolutional neural network, and the darker color represents the higher risk of nodule classification. In the left column (A), ultrasonogram shows a mixed echogenic thyroid nodule, in which the solid portion has irregular morphology, unclear demarcation from the surrounding parenchyma, and multiple internal microcalcifications, diagnosed as a C-TIRADS category 4 nodule, Figure 2A heat map shows areas of dense aggregation with warm colors; Figure 2B ultrasonogram shows an isoechoic thyroid nodule with irregular morphology, an aspect ratio of greater than 1, and posterior echogenic attenuation, diagnosed as a C-TIRADS category 4 nodule, Figure 2B heat map shows warm-toned color as an area of dense aggregation; Figure 2C ultrasonogram shows a hypoechoic nodule with regular morphology and homogeneous internal echogenicity, diagnosed as a C-TIRADS category 3 nodule, and Figure 2C heat map shows warm-toned color as an area.

3.5 Computational analysis

We further analyzed the computational costs of the compared methods by counting the averaging inference time of each patient. We listed the counted time in Table 5, where the results show that our model still retain outperforming computation efficiency when achieving the best accuracy. As our model only predict the probability once per patient rather than predict that once per image, the inference time of ours is lower than that of other models.

Table 5
www.frontiersin.org

Table 5. Results of our model and comparison methods.

4 Discussion and conclusion

4.1 Discussion

This study developed a multi-instance learning convolutional neural network model of ultrasound features to automatically recognize thyroid nodules and perform risk assessment. Ultrasound images of 2000 patients with thyroid nodules diagnosed by sonographers based on C-TIRADS were collected. The results of this experiment show that the multi-instance learning convolutional neural network model outperforms several metrics, especially in recall (0.8427) and AUC (0.8900) as shown in Table 2. Our model is important for clinical diagnosis, especially for identifying and classifying high-risk patients with high accuracy, which can effectively reduce the risk of missed diagnoses.

In clinical applications, when physician encounters a thyroid nodule with uncertain or controversial C-TIRADS grading, we import multiple thyroid nodule ultrasound images into the multi-instance learning convolutional neural network model, which extracts more subtle features from the images, providing diagnostic information that cannot be obtained by the sonographer’s naked eye, and assisting the sonographer to make a more accurate judgment. In the future, we will also integrate the model into the ultrasound workstation, so that after the sonographer acquires multiple ultrasound images of thyroid nodules, the system automatically inputs the images into the model, and the model quickly outputs the prediction results of the nodule classification, providing real-time references for the physician to help him or her to more accurately perform the C-TIRADS grading. In the future, the model will further learn a large amount of multi-center clinical data to quantitatively assess the risk of malignancy of thyroid nodules and give a specific risk value to make up for the relative subjectivity of the C-TIRADS classification and provide physicians with a more accurate diagnosis basis.

Compared with previous studies, our study offers several advantages. First, while the MIL model has been widely used in the preprocessing stage of pathology images, we extended it to ultrasound images for thyroid nodule risk assessment. In this study, we considered the entire set of ultrasound image features of the patient, accounting for differences in lesion presentation across multiple views. By aggregating these features using the MIL framework, we achieved a more comprehensive and robust assessment of the patient’s overall lesion condition. The MIL approach enables the model to analyze the entire set of ultrasound images (i.e., video frames), rather than relying on a single image, which mitigates the risks of misjudgment and annotation inconsistencies. Additionally, MIL allows for patient-level assessments, where the risk label of the ultrasound image set is based on the physician’s diagnosis using C-TIRADS, eliminating the need to label each individual image. This significantly reduces labeling time and improves accuracy, offering advantages over traditional supervised learning models (16). Table 2 compares our model with 13 typical deep learning models, including ResNet, AlexNet, VGGNet, MobileNet, and DenseNet, showing that our model achieves the best experimental results. In Table 2, comparisons with ResNet18, VGG16, DenseNet169, and DenseNet201 reveal no statistical differences between our model and these models. Moreover, VGG13 outperforms the other models in accuracy (0.8386), precision (0.8512), recall (0.8427), and F1-Score (0.8380), demonstrating the advantage of its deep convolutional structure in capturing thyroid ultrasound features. The highest F1-Score indicates a good balance between precision and recall, minimizing false positives while capturing malignant cases effectively. Statistical analysis in Table 2 supports the superiority of our model, showing significant differences compared to most models (p < 0.05). For instance, compared with ResNet34, ResNet50, AlexNet, and DenseNet161, our model not only exhibits statistically significant differences but also demonstrates substantial improvements with large effect sizes (Cohen’s d > 0.3). Additionally, the differences between our model and VGG11, the MobileNet series, and DenseNet121 are statistically significant (p < 0.0001). However, their effect sizes are smaller (Cohen’s d < 0.2), indicating moderate to minor practical differences. In cases where statistically non-significant results were observed, they may still hold value in the broader context of this study. For example, when comparing our model with ResNet18, VGG16, DenseNet169, and DenseNet201, although no statistically significant differences were identified, the observed performance metrics consistently favor our model. This consistency suggests that our approach is at least comparable, if not superior, in these scenarios. Moreover, the lack of statistical significance could be attributed to the relatively small effect sizes (Cohen’s d < 0.2) or the inherent variability within the dataset, rather than a true lack of difference. When compared with the baseline methods, our model demonstrated statistical significance, with effect sizes of 0.3762 and 0.1824, respectively. Our model has satisfactory classification performance, is more reliable in identifying high-risk thyroid patients, effectively reduces the leakage rate, and is important in clinical diagnosis, helping improve physicians’ ability to recognize the benign and malignant nature of thyroid nodules.

The model provides physicians with objective risk assessments by extracting aggregated features from thyroid nodule ultrasound images. This helps doctors make more informed treatment decisions, combining these results with clinical history and other test outcomes. Accurate risk assessment is crucial for determining appropriate treatment options. The model reduces unnecessary invasive tests and treatments while ensuring timely intervention for high-risk nodules. By automating feature extraction and risk assessment, the model enhances the efficiency of medical treatment, particularly in areas with a high prevalence of thyroid nodules, and significantly improves the accuracy and efficiency of ultrasound examinations.

4.2 Limitations

This study has several limitations. First, demographic information was not collected at the time of data collection for deeper sensitivity analysis. In future studies, baseline information on patients and populations receiving FNA for thyroid nodules should be collected to increase the diversity of available data. Second, single data modality was used in this study. The absence of such multi-modal data limits the model’s ability to provide a comprehensive risk assessment. Integrating these additional data sources in future studies could enhance diagnostic accuracy and personalization. Third, all data used in this study were collected from a single medical center, raising concerns about the generalizability of our findings. Variations in imaging protocols, equipment, and patient populations across institutions may affect model performance. To address this limitation, we plan to conduct cross-center validation and external testing in future work to evaluate the model’s robustness and adaptability in diverse clinical environments. Fourth, obtaining large-scale, high-quality labeled ultrasound datasets remains a significant challenge. Manual annotation by expert radiologists is time-consuming, costly, and subject to inter-observer variability. These factors can limit the scalability and consistency of training data. Future research should explore weakly supervised or semi-supervised learning approaches to reduce reliance on extensive manual labeling. Fifth, data annotation was performed by a single expert radiologist, which may introduce labeling bias. Future work should include multiple annotatorsm and consensus-based labeling for ambiguous cases. Although this study shows potential for assessing the risk of thyroid nodules, further research and improvements are needed to increase its validity and reliability in practical clinical applications.

4.3 Conclusion

In conclusion, the proposed multi-instance learning convolutional neural network model enables accurate and objective identification of high-risk thyroid nodules using multiple ultrasound images. Unlike traditional methods that rely heavily on subjective interpretation, our model leverages deep learning and attention mechanisms to automatically classify, localize, and assess the malignancy risk of thyroid nodules. To support this study, we constructed a real thyroid ultrasound dataset and performed a comprehensive comparison with existing thyroid nodule classification models using ultrasound images. The experimental results show that our convolutional classification model based on multi-instance learning exhibits an excellent performance. In summary, our innovative approach not only provides a new direction for expanding the application of multi-instance learning in whole-slice images (WSI) but also provides an automated C-TIRADS-based solution for thyroid nodules risk assessment.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by The Research Ethics Committee of the Affiliated Hospital of Hangzhou Normal University/The Affiliated Hospital of Hangzhou Normal University. The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because This retrospective study only analyzed data from previous ultrasound imaging data and medical records, and did not involve any invasive operations or interventions that would cause direct or indirect physical or psychological harm to the subjects, or disclose sensitive information that could identify the subjects, and based on the consideration of the risk-benefit ratio, the application was made for a waiver of written informed consent. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

DY: Validation, Formal Analysis, Writing – original draft, Data curation, Methodology, Investigation, Conceptualization, Visualization, Project administration. TS: Writing – original draft, Investigation, Supervision. YY: Data curation, Writing – original draft, Investigation. HZ: Validation, Project administration, Investigation, Writing – original draft. FG: Validation, Writing – review & editing, Supervision, Project administration. JW: Writing – original draft, Software, Visualization, Data curation, Validation. ZW: Data curation, Resources, Methodology, Project administration, Writing – review & editing, Supervision, Software, Funding acquisition.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This research was funded by the Medical and Health Science & Technology Project of Zhejiang Province, grant number 2023KY954, the Medical and Health Science & Technology Project of Zhejiang Province, grant number 2020KY712.

Acknowledgments

We acknowledge the participants for their consent to take part in the study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. (2021) 71:209–49. doi: 10.3322/caac.21660

PubMed Abstract | Crossref Full Text | Google Scholar

2. Pizzato M, Li M, Vignat J, Laversanne M, Singh D, La Vecchia C, et al. The epidemiological landscape of thyroid cancer worldwide: GLOBOCAN estimates for incidence and mortality rates in 2020. Lancet Diabetes endocrinology. (2022) 10:264–72. doi: 10.1016/S2213-8587(22)00035-3

PubMed Abstract | Crossref Full Text | Google Scholar

3. LeClair K, Bell KJL, Furuya-Kanamori L, Doi SA, Francis DO, and Davies L. Evaluation of gender inequity in thyroid cancer diagnosis: differences by sex in US thyroid cancer incidence compared with a meta-analysis of subclinical thyroid cancer rates at autopsy. JAMA Internal Med. (2021) 181:1351–8. doi: 10.1001/jamainternmed.2021.4804

PubMed Abstract | Crossref Full Text | Google Scholar

4. Wang B, Xi H, Wan Z, Huang S, Lai S, Miao X, et al. Analysis of epidemiology and clinical pathological characteristics of 9662 cases of thyroid cancer. Chin J Endocrine Surgery. (2021) 15:342–7. doi: 10.3760/cma.j.cn.115807‐20210609‐00178

Crossref Full Text | Google Scholar

5. Zhai M, Zhang D, Long J, Gong Y, Ye F, Liu S, et al. The global burden of thyroid cancer and its attributable risk factor in 195 countries and territories: A systematic analysis for the global burden of disease study. Cancer Med. (2021) 10:4542–54. doi: 10.1002/cam4.3970

PubMed Abstract | Crossref Full Text | Google Scholar

6. Deng Y, Li H, Wang M, Li N, Tian T, Wu Y, et al. Global burden of thyroid cancer from 1990 to 2017. JAMA Netw Open. (2020) 3:e208759. doi: 10.1001/jamanetworkopen.2020.8759

PubMed Abstract | Crossref Full Text | Google Scholar

7. Wang YQ, Li HZ, Gong WW, Chen YY, Zhu C, Wang L, et al. Cancer incidence and mortality in Zhejiang Province, Southeast China, 2016: a population-based study. Chin Med J (Engl). (2021) 134:1959–66. doi: 10.1097/CM9.0000000000001666

PubMed Abstract | Crossref Full Text | Google Scholar

8. Zheng RS, Sun KX, Zhang SW, Zeng HM, Zou XN, Chen R, et al. Analysis of Malignant tumor prevalence in China in 2015. Chin J Oncol. (2019) 41:19–28. doi: 10.3760/cma.j.issn.0253-3766.2019.01.005

PubMed Abstract | Crossref Full Text | Google Scholar

9. Zhou J, Yin L, Wei X, Zhang S, Song Y, Luo B, et al. 2020 Chinese guidelines for ultrasound Malignancy risk stratification of thyroid nodules: the C-TIRADS. Endocrine. (2020) 70:256–79. doi: 10.1007/s12020-020-02441-y

PubMed Abstract | Crossref Full Text | Google Scholar

10. Peng JY, Pan FS, Wang W, Wang Z, Shan QY, Lin JH, et al. Malignancy risk stratification and FNA recommendations for thyroid nodules: A comparison of ACR TI-RADS, AACE/ACE/AME and ATA guidelines. Am J Otolaryngol. (2020) 41:102625. doi: 10.1016/j.amjoto.2020.102625

PubMed Abstract | Crossref Full Text | Google Scholar

11. Jin Z, Pei S, Shen H, Ouyang L, Zhang L, Mo X, et al. Comparative study of C-TIRADS, ACR-TIRADS, and EU-TIRADS for diagnosis and management of thyroid nodules. Acad Radiol. (2023) 30:2181–91. doi: 10.1016/j.acra.2023.04.013

PubMed Abstract | Crossref Full Text | Google Scholar

12. Ito Y, Miyauchi A, and Oda H. Low-risk papillary microcarcinoma of the thyroid: A review of active surveillance trials. Eur J Surg Oncol. (2018) 44:307–15. doi: 10.1016/j.ejso.2017.03.004

PubMed Abstract | Crossref Full Text | Google Scholar

13. Ha EJ, Na DG, Baek JH, Sung JY, Kim JH, and Kang SY. US fine-needle aspiration biopsy for thyroid Malignancy: diagnostic performance of seven society guidelines applied to 2000 thyroid nodules. Radiology. (2018) 287:893–900. doi: 10.1148/radiol.2018171074

PubMed Abstract | Crossref Full Text | Google Scholar

14. Alyami J, Almutairi FF, Aldoassary S, Albeshry A, Almontashri A, Abounassif M, et al. Interobserver variability in ultrasound assessment of thyroid nodules. Med (Baltimore). (2022) 101:e31106. doi: 10.1097/MD.0000000000031106

PubMed Abstract | Crossref Full Text | Google Scholar

15. Persichetti A, Di Stasio E, Coccaro C, Graziano F, Bianchini A, Di Donna V, et al. Inter- and intraobserver agreement in the assessment of thyroid nodule ultrasound features and classification systems: A blinded multicenter study. Thyroid. (2020) 30:237–42. doi: 10.1089/thy.2019.0360

PubMed Abstract | Crossref Full Text | Google Scholar

16. Allahem H, El-Ghany SA, Abd El-Aziz AA, Aldughayfiq B, Alshammeri M, and Alamri M. A Hybrid model of feature extraction and dimensionality reduction using ViT, PCA, and random forest for multi-classification of brain cancer. Diagnostics (Basel). (2025) 15:1392. doi: 10.3390/diagnostics15111392

PubMed Abstract | Crossref Full Text | Google Scholar

17. Elhadidy MS, Elgohr AT, El-Geneedy M, Akram S, and Kasem HM. Comparative analysis for accurate multi-classification of brain tumor based on significant deep learning models. Comput Biol Med. (2025) 188:109872. doi: 10.1016/j.compbiomed.2025.109872

PubMed Abstract | Crossref Full Text | Google Scholar

18. Bai Z, Chang L, Yu R, Li X, Wei X, Yu M, et al. Thyroid nodules risk stratification through deep learning based on ultrasound images. Med Phys. (2020) 47:6355–65. doi: 10.1002/mp.14543

PubMed Abstract | Crossref Full Text | Google Scholar

19. Pan L, Cai Y, Lin N, Yang L, Zheng S, and Huang L. A two-stage network with prior knowledge guidance for medullary thyroid carcinoma recognition in ultrasound images. Med physics. (2022) 49:2413–26. doi: 10.1002/mp.15492

PubMed Abstract | Crossref Full Text | Google Scholar

20. Zhao Z, Yang C, Wang Q, Zhang H, Shi L, and Zhang Z. A deep learning-based method for detecting and classifying the ultrasound images of suspicious thyroid nodules. Med Phys. (2021) 48:7959–70. doi: 10.1002/mp.15319

PubMed Abstract | Crossref Full Text | Google Scholar

21. Sun C, Zhang Y, Chang Q, Liu T, Zhang S, Wang X, et al. Evaluation of a deep learning-based computer-aided diagnosis system for distinguishing benign from Malignant thyroid nodules in ultrasound images. Med Phys. (2020) 47:3952–60. doi: 10.1002/mp.14301

PubMed Abstract | Crossref Full Text | Google Scholar

22. Li X, Zhang S, Zhang Q, Wei X, Pan Y, Zhao J, et al. Diagnosis of thyroid cancer using deep convolutional neural network models applied to sonographic images: a retrospective, multicohort, diagnostic study. Lancet Oncol. (2019) 20:193–201. doi: 10.1016/S1470-2045(18)30762-9

PubMed Abstract | Crossref Full Text | Google Scholar

23. Peng S, Liu Y, Lv W, Liu L, Zhou Q, Yang H, et al. Deep learning-based artificialintelligence model to assist thyroid nodule diagnosis and management: a multicentre diagnostic study. Lancet DigitHealth. (2021) 3:e250–9. doi: 10.1016/S2589-7500(21)00041-8

PubMed Abstract | Crossref Full Text | Google Scholar

24. Zhao SX, Chen Y, Yang KF, Luo Y, Ma BY, and Li YJ. A local and global feature disentangled network: toward classification of benign-malignant thyroid nodules from ultrasound image. IEEE Trans Med Imaging. (2022) 41:1497–509. doi: 10.1109/TMI.2022.3140797

PubMed Abstract | Crossref Full Text | Google Scholar

25. Brooks FJ and Grigsby PW. Quantification of heterogeneity observed in medical images. BMC Med imaging. (2013) 13:1–12. doi: 10.1186/1471-2342-13-7

PubMed Abstract | Crossref Full Text | Google Scholar

26. Sylolypavan A, Sleeman D, Wu H, and Sim M. The impact of inconsistent human annotations on AI driven clinical decision making. NPJ Digital Med. (2023) 6:26. doi: 10.1038/s41746-023-00773-3

PubMed Abstract | Crossref Full Text | Google Scholar

27. Xue LF, Zhang XL, Tang YF, and Wei BH. Multi-instance learning for identifying high-risk subregions associated with synchronous distant metastasis in clear cell renal cell carcinoma. Med Phys. (2024) 51:9115–24. doi: 10.1002/mp.17439

PubMed Abstract | Crossref Full Text | Google Scholar

28. Tian Y, Hao W, Jin D, Chen G, and Zou A. (2020). A review of latest multi-instance learning, in: Proceedings of the 2020 4th International Conference on Computer Science and Artificial Intelligence. New York: Association for Computing Machinery. pp. 41–5. doi: 10.1145/3445815.344582

Crossref Full Text | Google Scholar

29. Zhou ZH. Multi-instance learning: A survey. Nanjing, China: Department of Computer Science & Technology, Nanjing University, Tech. Rep (2004) p. 1–4.

Google Scholar

30. Carbonneau MA, Cheplygina V, Granger E, and Gagnon G. Multiple instance learning: A survey of problem characteristics and applications. Pattern Recognit. (2018) 77:329–53. doi: 10.1016/j.patcog.2017.10.009

Crossref Full Text | Google Scholar

31. Quellec G, Cazuguel G, Cochener B, and Lamard M. Multiple-instance learning for medical image and video analysis. IEEE Rev Biomed engineering. (2017) 10:213–34. doi: 10.1109/RBME.2017.2651164

PubMed Abstract | Crossref Full Text | Google Scholar

32. El-geneedy M, Elgohr AT, Elhadidy MS, and Akram S. Early lung cancer detection with a fusion of inception V3 and vision transformers: A binary classification study. In: 2024 International Conference on Future Telecommunications and Artificial Intelligence (IC-FTAI). Red Hook, NY USA: IEEE (2024). p. 1–6. doi: 10.1109/IC-FTAI62324.2024.10950031

Crossref Full Text | Google Scholar

33. Ilse M, Tomczak J, and Welling M. Attention-based deep multiple instance learning. In: International conference on machine learning. Massachusetts, USA: PMLR (2018). p. 2127–36. Available online at: https://arxiv.org/abs/1802.04712 (Accessed July 10-15, 2018).

Google Scholar

34. He K, Zhang X, Ren S, and Sun J. Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition. Red Hook, NY USA: CVPR (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

35. Krizhevsky A, Sutskever I, and Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems. New York, NY, United States: Association for Computing Machinery (2012). p. 1097–105.

Google Scholar

36. Simonyan K and Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv. (2014), 1409–556. Available online at: https://arxiv.org/abs/1409.1556 (Accessed April 10, 2015).

Google Scholar

37. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv. (2017), 1704.04861. doi: 10.48550/arXiv.1704.04861

Crossref Full Text | Google Scholar

38. Huang G, Liu Z, van der Maaten L, and Weinberger KQ. (2017). Densely connected convolutional networks, in: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).Red Hook, NY USA: IEEE. pp. 4700–8.

Google Scholar

39. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, and Batra D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int J Comput Vis. (2020) 128:336–59. doi: 10.1007/s11263-019-01228-7

Crossref Full Text | Google Scholar

Keywords: multi-instance learning, convolutional neural network, thyroid nodule, ultrasound image feature, thyroid cancer diagnosis

Citation: Yu D, Song T, Yu Y, Zhang H, Gao F, Wang Z and Wang J (2025) Risk assessment of thyroid nodules with a multi-instance convolutional neural network. Front. Oncol. 15:1608963. doi: 10.3389/fonc.2025.1608963

Received: 09 April 2025; Accepted: 01 July 2025;
Published: 24 July 2025.

Edited by:

Sharon R Pine, University of Colorado Anschutz Medical Campus, United States

Reviewed by:

Mohamed S. Elhadidy, Horus University, Egypt
Youness Khourdifi, Université Sultan Moulay Slimane, Morocco

Copyright © 2025 Yu, Song, Yu, Zhang, Gao, Wang and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zirong Wang, MjAxNzEwOTBAaHpudS5lZHUuY24=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.