Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Physiol., 13 November 2025

Sec. Computational Physiology and Medicine

Volume 16 - 2025 | https://doi.org/10.3389/fphys.2025.1689936

Clinically guided adaptive contrast adjustment for fetal plane classification: a modular plug-and-play solution

  • 1School of Mathematics and Statistics, Xiamen University of Technology, Xiamen, China
  • 2Hunan University of Finance and Economics, Changsha, China
  • 3Dongguan University of Technology, Dongguan, China
  • 4School of Business, Society and Engineering, Mälardalens University, Västerås, Sweden

Fetal ultrasound standard plane recognition plays a vital role in ensuring accurate prenatal assessment but remains challenging due to intrinsic factors such as poor tissue contrast, indistinct anatomical boundaries, and variability in image quality caused by operator differences. To address these issues, we introduce a plug-and-play Adaptive Contrast Adjustment Module (ACAM), inspired by how clinicians manually adjust image contrast to highlight clearer structural cues. The proposed module integrates a lightweight, texture-aware subnetwork that learns to generate clinically meaningful contrast parameters, producing multiple contrast-enhanced representations of the same image through a differentiable transformation process. These enhanced views are then fused within subsequent classifiers to enrich discriminative features. Experiments conducted on a multi-center dataset containing 12,400 fetal ultrasound images across six anatomical planes demonstrate consistent performance gains: the accuracy of lightweight models rises by 2.02%, conventional architectures by 1.29%, and state-of-the-art models by 1.15%. The key novelty of ACAM lies in its content-adaptive and clinically aligned contrast modulation, which replaces random preprocessing with physics-guided transformations mimicking sonographers’ diagnostic workflows. By leveraging multi-view contrast fusion, our approach enhances robustness against image quality variations and effectively links low-level texture cues with high-level semantic understanding, offering a new framework for medical image analysis in realistic clinical settings. Our code is available at: https://github.com/sysll/ACAM.

1 Introduction

Ultrasound offers several advantages, including safety, convenience, non-invasiveness, and the absence of ionizing radiation, which has led to its widespread application in critical areas such as prenatal fetal screening Maher and Seed (2024); Wittek et al. (2025); Al-Dahim et al. (2024); Miller et al. (2020); Wang (2018). The acquisition of standardized fetal ultrasound planes is essential for improving diagnostic precision and minimizing the risk of overlooking severe fetal abnormalities. Nevertheless, achieving this standardization remains challenging: it requires operators to have comprehensive knowledge of fetal anatomy, while clinical expertise and equipment conditions may sometimes be inadequate. Furthermore, the increasing complexity of screening settings, the rising demand for fetal examinations, and the shortage of skilled ultrasound practitioners make manual acquisition of high-quality planes even more difficult. In this context, there is a pressing need for automated recognition systems to support sonographers in efficiently and accurately identifying standard fetal trunk planes. These systems can help reduce missed diagnoses, improve workflow efficiency, and provide more reliable and safer technical assistance for prenatal evaluation.

Deep learning has demonstrated remarkable capabilities and has been widely applied across various domains Cai et al. (2024); Zhu et al. (2024); Ou et al. (2024); Mykula et al. (2024); Zhao et al. (2024a); Zhao et al. (2024b); Jin et al. (2025); Zhao et al. (2025). In recent years, there has been growing interest in algorithms for fetal ultrasound plane analysis Zhu et al. (2025b), Zhu et al. (2025a); Boumeridja et al. (2025); Montero et al. (2021); Yousefpour Shahrivar et al. (2023); Krishna and Kokil (2024); Krishna and Kokil (2023); Fiorentino et al. (2025a); Migliorelli et al. (2024). However, most studies primarily focus on feature extraction modules, emphasizing information in intermediate network layers or increasing dataset size to improve model performance. For example, Zhu et al. (2025b) aimed to optimize pooling layer performance; while insightful, this approach overlooks the impact of the input layer. Similarly, Montero et al. (2021) employed generative adversarial networks (GANs) to generate additional training images, thereby enlarging the dataset. Only a few studies consider the interaction between the model and the input image in relation to image quality. For instance, Zhu et al. (2025a) highlighted the importance of selecting appropriate contrast and gain for medical image performance and proposed an attention mechanism to focus on regions with critical gain. However, in their approach, contrast and gain are fixed rather than adaptively generated, which limits the model’s capability. To address these limitations, we propose an Adaptive Contrast Adjustment Module (ACAM) that dynamically adjusts image contrast based on image content. By generating multiple contrast-enhanced versions and fusing their information, the module not only enriches texture representations but also significantly improves the classification accuracy of complex fetal plane images.

Our approach is motivated by the practical workflow of clinicians when identifying fetal planes during ultrasound examinations. In routine practice, sonographers often manipulate image contrast to emphasize key anatomical structures, which helps produce clearer and more discriminative images Smith and Lopez (1982); Mehta et al. (2017). Drawing from this idea, we incorporate an adaptive contrast adjustment module into our model. Specifically, the input image is first processed by a decision network that predicts K potential contrast parameters. These predicted parameters are then mapped to a predefined fixed range using a differentiable function to ensure numerical stability and preserve trainability. Using these contrast parameters, the input image is transformed to produce K contrast-enhanced variants, effectively introducing multiple perspectives or styles during training. These enhanced images are then passed through a convolutional neural network for feature extraction and classification. Since contrast adjustment relies more on local texture information rather than high-level semantic cues, we employ a shallow convolutional network as the decision module. This network captures fine-grained details, such as edges and textures, to generate the contrast parameters. This design provides both interpretability and generalization benefits. The parameters directly control image brightness and contrast, making the transformations intuitive and visually interpretable, in contrast to black-box manipulations of abstract features. Additionally, by explicitly generating multiple contrast scenarios, the model learns representations that are more robust to variations in illumination and contrast, which improves generalization across different acquisition settings or imaging domains.

Furthermore, our module adopts a plug-and-play architecture and is applied solely to the lower layers of the network, allowing for easy integration. We evaluated its effectiveness by embedding it into conventional robust models, lightweight networks, and cutting-edge architectures, performing ablation studies to quantify its impact. Comparative experiments were subsequently conducted against eight baseline models. The results indicate that incorporating our module consistently improves performance. The main benefits of the module are summarized as follows:

• The module emulates the way clinicians adjust image contrast, allowing adaptive generation of multiple images with different contrast levels. This enables the model to learn from diverse representations, enhancing its sensitivity to fine details and improving overall robustness.

• In our framework, a shallow convolutional network first extracts local texture information from the input image. Using these features, the network predicts several candidate contrast values, which are then applied to enhance the image and enrich feature representation.

• We incorporated the module into lightweight CNNs, conventional robust models, and state-of-the-art architectures, performing comprehensive evaluations. Comparative experiments, ablation studies, and heatmap visualizations confirm that the module consistently boosts model performance and generalizability.

2 Methods

2.1 Linear contrast

Image contrast enhancement can be achieved through either linear or nonlinear gray-level transformations, with the basic goal of stretching or compressing the distribution range of pixel intensities, thereby emphasizing the intensity differences across regions of the image. Let the original grayscale image be denoted as I(x,y), where (x,y) represents the pixel location in the image. A commonly used linear contrast adjustment method can be formulated as shown in Equation 1:

Ix,y=αIx,yμ+μ,(1)

where I(x,y) represents the adjusted pixel intensity, α>0 is the contrast scaling factor (typically referred to as the contrast gain), and μ is the mean intensity (brightness center) of the image, defined as shown in Equation 2:

μ=1HWx=1Hy=1WIx,y,(2)

with H and W denoting the image height and width, respectively.

When α>1, the contrast of the image is increased, whereas α<1 reduces the contrast.

2.2 The mechanism of the ACAM module

The structure of our module is illustrated in Figure 1. First, the input is a grayscale image with dimensions [1, H, W]. The first step of the model is to generate a set of contrast values from this image for subsequent processing. We posit that contrast prediction primarily relies on the detailed information within the image rather than semantic-level features. Therefore, this module employs a shallow architecture composed of convolutional layers, a global average pooling layer, and fully connected layers. This design is chosen because shallow convolutional neural networks are more adept at extracting high-frequency detail information from images, whereas deeper convolutions mainly capture semantic features. Moreover, using a low-level structure introduces fewer parameters. The predicted contrast values are then mapped to the range [1, 3] to align with the adjustment range typically used by clinicians. This process can be expressed as Equation 3.

C=FunctionFCGAPConvImage(3)

where C=[c1,c2,,ck]. The Function(x) function is shown in Equation 4.

Functionx=1+211+ex(4)

Figure 1
Diagram illustrating a process for ultrasound image analysis using a convolutional neural network (CNN). The left panel shows an ultrasound image input into the CNN, which generates contrast variations represented by grid-like structures. These variations undergo contrast adjustment, yielding processed images. These images are then fed into a classifier for further analysis. The workflow indicates steps for improving image contrast and classification.

Figure 1. Architecture of the proposed module. It consists of two components: contrast generation and contrast adjustment. The contrast generation component predicts k distinct contrast parameters, which are subsequently used by the contrast adjustment component to transform the input image accordingly.

These contrast values are then fed into the model as contrast parameters. The specific formula is given in Equation 5.

Ikij=ckIkij1HWijIkij+1HWijIkij(5)

Where Ikij denotes the pixel at row i, column j of the k-th image. The final output (I) is with dimensions of [K×H×W], representing k images generated under k different contrast conditions. These images constitute the output of our module and serve as input to subsequent decision-making models such as MedMamba. Clinicians typically begin by adjusting image contrast to enhance clarity before proceeding to in-depth analysis for diagnostic classification. Our ACAM module mimics this contrast enhancement process to optimize visual clarity, operating primarily on low-level texture information. In contrast, downstream decision-making models such as MedMamba simulate the clinician’s diagnostic reasoning process, which necessitates the extraction of high-level semantic features, so they have deeper layers.

2.3 Implementation details

This study is based on a large-scale prenatal screening ultrasound image dataset Burgos-Artizzu et al. (2020), which was collected from two hospitals and encompasses multiple operators as well as different ultrasound device models. All images were manually annotated by a single obstetrics expert and categorized into six classes: four commonly used fetal standard planes (abdominal, brain, femur, and thoracic), the maternal cervix plane for preterm screening, and a general class including other less common planes. The names of these standard planes and their corresponding encoded categories are shown in Figure 2. The number of images for each standard plane category is shown in the Table 1. The final dataset comprises over 12,400 images from 1,792 patients, and it was split into training and test sets at a ratio of 7:3. All experiments were conducted using Python 3.9 and the PyTorch 2.0.1+cu117 framework, on a system equipped with an Intel i7-12650H processor and an NVIDIA RTX2080Ti GPU. Detailed settings of the model parameters and baseline models are provided in Table 2, with the number of generated contrast images n set to 10.

Figure 2
Six ultrasound images labeled with classifications:

Figure 2. Sample images from each class of the dataset.

Table 1
www.frontiersin.org

Table 1. The number of each fetal ultrasound standard plane in the dataset.

Table 2
www.frontiersin.org

Table 2. Hyperparameter settings used during model training.

We justify our hyperparameter choices as follows. The contrast range is set to [1,3], based on clinical practice. The study Zhu et al. (2025c) discussed the changes in ultrasound images with contrast ranging from 1 to 2.5, illustrating the advantages of this range. Clinicians typically adjust image contrast within this range to enhance the visibility of anatomical structures while avoiding image distortion—values below 1.0 excessively compress the dynamic range, whereas values above 3.0 may cause key regions to become oversaturated. The number of contrast variants is configured at K=10, taking into account clinical practices. While clinicians may experiment with various contrast adjustments for a single image, they seldom employ an excessive number. In the study by Zhu et al. (2025a), nine contrast groups with fixed gain settings were used, which is close to our setting. Therefore, choosing 10 variants strikes a balance between computational efficiency and feature diversity. For the network architecture, we use a shallow convolutional network for contrast prediction to extract texture features, as contrast adjustment mainly relies on low-level image statistics rather than high-level semantic information. Studies Zhang et al. (2019); Gao et al. (2020); Cimpoi et al. (2015) have all highlighted the role of shallow neural networks in extracting texture information. In terms of training configuration, the batch size is set to 64 to efficiently utilize GPU memory and follow the powers-of-two convention; the learning rate is set to 0.001 to ensure stable convergence; and the model is trained for 20 epochs to achieve sufficient convergence while avoiding overfitting. This setup is based on the experience reported in previous studies on related datasets Burgos-Artizzu et al. (2020); Zhu et al. (2025a); Montero et al. (2021).

3 Results

3.1 Evaluation metrics and baseline models

In this study, multiple widely adopted evaluation metrics are employed to systematically analyze model performance. Accuracy (ACC) reflects the overall correctness of predictions; however, it may be misleading in scenarios with imbalanced class distributions. Recall measures the model’s ability to correctly identify positive samples, which is particularly crucial in medical image analysis, as higher recall helps reduce the risk of missed diagnoses. Precision evaluates the proportion of predicted positive samples that are truly positive, thereby reducing the likelihood of false alarms. The F1-score, defined as the harmonic mean of precision and recall, provides a balanced assessment of both metrics.

Here, we denote the standard confusion matrix terms as follows:

TP (True Positive): number of correctly predicted positive samples,

TN (True Negative): number of correctly predicted negative samples,

FP (False Positive): number of negative samples incorrectly predicted as positive,

FN (False Negative): number of positive samples incorrectly predicted as negative.

Based on these definitions, the metrics are computed as Equations 69:

Accuracy=TP+TNTP+TN+FP+FN(6)
Precision=TPTP+FP(7)
Recall=TPTP+FN(8)
F1=2×Precision×RecallPrecision+Recall(9)

In addition, to comprehensively characterize the model’s classification capability across different decision thresholds, we introduce the Receiver Operating Characteristic (ROC) curve and employ the Area Under the Curve (AUC) as a performance indicator. Similarly, the Precision–Recall (PR) curve is utilized to illustrate prediction accuracy at varying recall levels, with the Average Precision (AP) computed to intuitively reflect the model’s ability in target detection tasks.

To evaluate the effectiveness of the proposed model, we compare it against several established deep learning architectures, including EfficientNet Kashyap et al. (2023), InceptionV3 Szegedy et al. (2016), VGG Gunasekaran and Vivekasaran (2024), ResNet Xu et al. (2023), MobileNet Han et al. (2022), ShuffleNet Hou et al. (2025), ConvNeXt Sangeetha and Geetha (2024), MedMamba Bansal et al. (2024), EfficientVMamba Pei et al. (2025), OrthoNets Salman et al. (2023), and Efficientvit Liu et al. (2023).

3.2 Comparison experiment

The performance comparison of the models is presented in Table 3. As shown, all evaluated models—ranging from lightweight networks such as ShuffleNet, MobileNet, and EfficientNet to traditional robust architectures including ResNet, VGG, InceptionV3, and ConVNeXt, as well as state-of-the-art deep learning models such as MedMamba variants and EfficientViT—achieved strong performance on the test set, with overall accuracy consistently exceeding the 90% baseline. Specifically, classical architectures like EfficientNet and InceptionV3 achieved top-1 accuracies of 92.26% and 92.32%, respectively, while MobileNet and VGG attained slightly lower accuracies of 90.27% and 90.73%. In addition to these baseline models, we evaluated RCJ-based models, which use Random Contrast Jittering as a data augmentation strategy. The incorporation of RCJ generally led to modest improvements across different backbones. For instance, RCJ-ResNet improved the accuracy from 91.72% to 92.02%, RCJ-MedMamba increased from 92.32% to 92.48%, and RCJ-ShuffleNet improved from 89.28% to 89.39%. These results indicate that contrast-based augmentation contributes to better robustness against intensity variations. The proposed ACAM module (Adaptive Contrast Adjustment Module), when integrated into different backbone networks, consistently improved model performance. ACAM-MedMamba achieved the highest accuracy of 93.47% and an F1-score of 93.47%, surpassing both the original MedMamba (92.32% accuracy, 92.36% F1-score) and RCJ-MedMamba (92.48% accuracy, 92.53% F1-score). Similarly, ACAM-ResNet improved accuracy from 91.72% to 93.01%, and ACAM-ShuffleNet increased accuracy from 89.28% to 91.30%. These results demonstrate the generalization capability of the ACAM module across different architectures. Overall, Table 3 shows that ACAM not only outperforms baseline and RCJ-enhanced models but also effectively enhances feature discrimination and complements existing data augmentation strategies, providing a robust approach for medical image classification tasks.

Table 3
www.frontiersin.org

Table 3. Ablation study results of our module integrated into different models, as well as comparisons with other models.

3.3 Ablation study

The results of the ablation study are summarized in Table 3. It can be seen that, regardless of whether the backbone is a traditional model (ResNet), a lightweight model (ShuffleNet), or a state-of-the-art model (MedMamba), integrating the proposed module leads to a significant performance improvement, with an average gain of 1.48%. This consistent enhancement across different architectures demonstrates the effectiveness and generality of the proposed module.

A comparison of confusion matrices, as shown in Figure 3, reveals that the ACAM module consistently improves classification performance across lightweight models (ShuffleNet), traditional models (ResNet), and state-of-the-art models (MedMamba). In particular, the classification accuracy for classes 0 and 1 is significantly enhanced in all models, with a substantial reduction in misclassifications. For class 5, most cases also show improved precision after module integration. These results highlight that ACAM can robustly optimize feature discrimination for both common and challenging classes across various backbone networks. Furthermore, the module effectively mitigates inter-class confusion, especially in models prone to overfitting or with limited representational capacity, confirming its generalization and robustness.

Figure 3
Six confusion matrices compare different models: Medmamba, ResNet, ShuffleNet, ACAM-Medmamba, ACAM-ResNet, and ACAM-ShuffleNet. Each matrix displays true labels versus predicted labels for classes 0 to 5, indicating varying prediction accuracies. The diagonal values represent correct predictions, with other values showing misclassifications. The matrices illustrate comparative performance, with darker blue indicating higher true positive counts.

Figure 3. Comparison of confusion matrices for three models before and after integrating the proposed module. (a–c) show the classification performance of the baseline models, while (d–f) illustrate the improvements achieved after incorporating the module.

As shown in Figure 4, the ROC curve analysis demonstrates that integrating the ACAM module significantly improves the classification performance of various models. Across the lightweight ShuffleNet, the conventional ResNet, and the advanced MedMamba, the trade-off between true positive rate (TPR) and false positive rate (FPR) is markedly enhanced for most classes. Specifically, after incorporating ACAM, MedMamba achieves notable AUC improvements of approximately 3%, 2%, and 2% for classes 0, 3, and 5, respectively; ResNet shows clear AUC gains of about 10%, 7%, and 1% for classes 0, 2, and 3; while ShuffleNet also exhibits appreciable AUC improvements of around 10% and 4% for classes 0 and 2. These observations further validate that ACAM provides consistent AUC enhancement and robustness across different model architectures. The precision–recall (PR) curves, shown in Figure 5, indicate that the module significantly enhances classification performance for key classes. In ShuffleNet, ACAM effectively improves the balance between precision and recall for classes 0, 1, and 2, with AP increases of approximately 15%, 2%, and 6%, respectively. For ResNet, notable improvements are observed in classes 0, 1, and 2, with AP increases of about 18%, 2%, and 12%. In MedMamba, classes 0, 3, and 5 clearly benefit from the module, with AP increases of roughly 3%, 3%, and 2%. These results suggest that ACAM can adaptively enhance the recognition of challenging samples according to the characteristics of different backbone networks, achieving higher recall while maintaining high precision, thereby demonstrating its broad applicability and effectiveness in improving classification performance.

Figure 4
Six ROC curve graphs comparing different models: (a) Medmamba, (b) ResNet, (c) ShuffleNet, (d) ACAM-Medmamba, (e) ACAM-ResNet, and (f) ACAM-ShuffleNet. Each graph displays true positive rate against false positive rate for six classes, providing Area Under Curve (AUC) values for each. Dotted diagonal represents random guessing. AUC values indicate varying model performance across classes, with ACAM models generally achieving higher AUC scores.

Figure 4. Comparison of ROC curves for three models before and after integrating the proposed module. (a–c) depict the classification performance of the baseline models, while (d–f) demonstrate the improvements achieved after incorporating the module.

Figure 5
Six precision-recall curve charts for different models: (a) Medmamba, (b) ResNet, (c) ShuffleNet, (d) ACAM-Medmamba, (e) ACAM-ResNet, and (f) ACAM-ShuffleNet. Each chart contains curves for six classes with specific average precision values. Charts compare model performance, displaying recall on the x-axis and precision on the y-axis.

Figure 5. Comparison of PR curves for three models before and after integrating the proposed module. (a–c) Illustrate the classification performance of the baseline models, while (d–f) demonstrate the improvements obtained after incorporating the module.

3.4 Heatmap-based and t-SNE visualization and analysis of detailed classification results

To further assess the effectiveness of the proposed ACAM module, we utilized the Grad-CAM technique to visualize the model’s attention regions. It should be noted that the visualizations are primarily based on ResNet, because Grad-CAM depends on the spatial feature maps of convolutional layers, which allow the generation of heatmaps with improved spatial alignment and interpretability in convolutional networks. As illustrated in Figure 6, the first column displays the original ultrasound images, while the second and third columns show the heatmaps produced by the baseline ResNet and the ACAM-enhanced ResNet (ACAM-ResNet), respectively. The results suggest that, unlike the baseline ResNet where attention areas are often scattered or misaligned with the relevant anatomical structures, ACAM-ResNet can concentrate more precisely on clinically important regions. For fetal thoracic planes, the baseline ResNet tends to distribute attention broadly across the thoracic cavity, whereas ACAM-ResNet significantly improves focus on critical organs, such as the heart and lungs. In the fetal femur planes, the baseline model may assign attention to surrounding soft tissues, but the ACAM-enhanced network accurately highlights the femoral shaft. In abdominal plane analysis, ACAM-ResNet shows more distinct attention toward structures such as the stomach bubble and umbilical cord insertion point, whereas the heatmaps from the baseline model are often diffuse. For fetal brain planes, the enhanced model clearly targets the lateral ventricles and midline structures, avoiding distraction from irrelevant brain regions. Moreover, in maternal cervical planes, ACAM-ResNet effectively emphasizes the internal cervical os and the cervical lumen, while the baseline model is easily diverted by adjacent tissues.

Figure 6
Ultrasound images of fetal anatomy and maternal cervix are shown in the left column, with corresponding heat maps from ResNet and ACAM-ResNet models in the middle and right columns, respectively. The heat maps highlight regions of interest in different colors, indicating attention and focus areas related to each anatomical site.

Figure 6. Heatmap visualizations of the ResNet model before and after integrating our module, illustrated on five representative image categories.

To analyze the feature distribution and inter-class relationships learned by different models, we plotted the t-SNE visualizations as shown in Figure 7. Analysis of the t-SNE visualization reveals that the feature clusters corresponding to the fetal brain and femur categories exhibit the most distinct separation, demonstrating clear isolation from other categories in the embedded space. With the exception of the “Other” category, all remaining classes maintain reasonably well-defined spatial boundaries. In the baseline ResNet model prior to integrating our ACAM module, feature representations of different categories appear in closer proximity, with substantial overlap observed particularly between the fetal thorax and “Other” categories. Following the incorporation of the ACAM module, the feature distributions show noticeable improvement in category separation, as evidenced by the more dispersed spatial arrangement of clusters. This observed expansion in inter-class distances demonstrates the module’s effectiveness in enhancing feature discriminability.

Figure 7
Side-by-side scatter plots comparing data separation using ResNet and ACAM-ResNet algorithms. Each plot features clusters differentiated by colors representing fetal abdomen, fetal femur, maternal cervix, fetal thorax, fetal brain, and other categories. The ResNet plot shows less distinct separation between clusters compared to the ACAM-ResNet, where the clusters are more distinct. A legend indicates color assignments for each category.

Figure 7. t-SNE visualization of feature embeddings extracted by ResNet and ACAM-ResNet models.

As shown in Table 4, the model performs well on most standard fetal planes, with the highest recognition achieved for the fetal femur (F1 = 0.9764) and fetal brain (F1 = 0.9869). This can be attributed to the distinctive anatomical features of these regions—specifically, the femur’s linear hyperechoic structure and the brain’s midline pattern—which provide stable cues for the model’s discrimination. However, the recall for the maternal cervix plane is relatively low (0.7442), primarily due to the following factors: first, the cervix exhibits significant morphological variation across different gestational weeks, ranging from a cylindrical to a funnel shape, resulting in large intra-class differences; second, even slight deviations in the probe angle can lead to incomplete visualization of the endometrial line, causing some positive samples to lack critical discriminative features; additionally, acoustic artifacts from the cervix plane overlapping with parts of the vaginal fornix introduce feature confusion. The precision for the fetal abdomen plane is also relatively low (0.8224), mainly because the abdominal plane often contains multiple solid organs (e.g., liver, intestines) with mixed echogenic patterns, which vary considerably across gestational ages and fetal positions. In particular, when the fetal abdomen includes amniotic fluid regions, it can be acoustically confused with fluid-filled structures in the thoracic cavity. Despite these challenges, the model maintains stable performance on most standard planes, demonstrating its ability to handle the inherent variability in fetal ultrasound images. Future work will incorporate attention mechanisms and domain adaptation strategies to further enhance the model’s discriminative capability on difficult samples.

Table
www.frontiersin.org

Table. The detailed performance of our integrated model on each fetal plane category based on MedMamba.

4 Discussion

4.1 Module significance and comparison with existing methods

In fetal ultrasound standard plane recognition, most methods rely on CNNs for texture and edge feature extraction Venkatareddy et al. (2024); Diniz et al. (2020); Wang et al. (2021), assuming input images of stable quality and moderate contrast. In clinical practice, however, factors such as fetal position, gestational age, device settings, and operator habits often cause substantial contrast variations, obscuring critical anatomical details. Clinicians typically adjust contrast to highlight essential structures, inspiring the design of our ACAM. Unlike conventional data augmentation, which applies random transformations without adapting to image content, ACAM dynamically models contrast in a content-aware manner, enhancing texture details and exploring multiple contrast perspectives. This approach preserves discriminative capability even with blurred structures or low signal-to-noise ratios. Beyond technical improvement for plane classification, ACAM reflects a paradigm aligning deep learning with clinical imaging practices, offering insights into medical AI by modeling contrast—a low-level yet clinically significant attribute.

Krishna and Kokil (2024) employed a stacked ensemble approach using three pre-trained deep CNNs: AlexNet, VGG-19, and DarkNet-19. Predictions from these networks were obtained via Softmax and random forest classifiers. In Krishna and Kokil (2023), AlexNet and VGG-19 were used to extract deep features, with a global average pooling layer as the final pooling layer for feature integration. Fusing deep features extracted from different convolutional networks enhances the overall feature representation. In contrast to their studies, which primarily focus on the diversity of extracted features, our work emphasizes adaptive adjustment of image contrast to improve image quality. Moreover, Venkatareddy et al. (2024) introduced explainable AI (XAI) methods—specifically Local Interpretable Model-agnostic Explanations (LIME)—to increase the transparency and reliability of model decisions. Our approach, however, introduces adaptive contrast generation, which not only enhances model performance but also improves the interpretability of the model design.

4.2 Secondary training strategy

Our model further supports an extended application. Specifically, the system can record clinicians’ contrast adjustment operations across various fetal ultrasound planes and use these records to supervise the training of the convolutional module in the contrast generation stage (Stage 1 in Figure 8). In the subsequent classification stage, the parameters of the first convolutional layer are frozen (Stage 2 in Figure 8). The core design of ACAM intrinsically simulates the clinical decision-making process: clinicians first adjust image contrast until the plane becomes sufficiently clear, and only then proceed with diagnosis. Our two-stage strategy closely aligns with this workflow by decomposing the task into two sequential objectives—first training the model to predict contrast, and then training the classification model using the contrast-enhanced images. This staged training paradigm not only improves model performance but also enhances interpretability, as the feature generation process explicitly reflects clinicians’ operational preferences. Furthermore, the method demonstrates strong extensibility, allowing adaptation to data acquired from different devices or operators, thereby further improving robustness and clinical applicability.

Figure 8
Diagram of a two-stage CNN model training process. In the first stage, a single-layer CNN model is trained using doctors' adjustment records. The CNN processes medical images and outputs classes \(C_1, C_2, \ldots, C_n\) for doctor validation. In the second stage, the CNN is frozen, and a classification model is trained using the frozen CNN output. Colors indicate training (pink) and trained models (blue).

Figure 8. Two-stage training strategy of ACAM, contrast generation (Stage 1) and classification (Stage 2).

4.3 Limitations and future directions

Although our method can automatically generate multiple contrast values from input images—enhancing the model’s sensitivity to fine-grained details—the number of generated contrast values is currently fixed. This design may limit adaptability when the model encounters extreme or previously unseen contrast variations. Future work could explore more flexible contrast generation mechanisms, such as variable-size or continuously parameterized approaches, to better capture a wider spectrum of contrast distributions and further improve robustness and generalization. Incorporating clinician adjustment records or prior clinical knowledge also represents a promising direction to enhance interpretability and clinical relevance.

Moreover, while our study demonstrates the effectiveness of ACAM on a widely used public fetal ultrasound benchmark, we acknowledge that relying on a single dataset may restrict generalizability. As highlighted by Fiorentino et al. (2025b), this dataset contains several biases, including class imbalance, demographic underrepresentation, and acquisition heterogeneity. These factors can affect model performance and may not fully reflect clinical variability in broader populations. By explicitly addressing these challenges, our work underscores the value of modules like ACAM in improving model robustness to image-level variations. Future studies will aim to validate ACAM on more diverse clinical datasets to further assess its generalizability and practical applicability in real-world settings.

5 Conclusion

This work presents ACAM, a novel paradigm for fetal ultrasound plane classification that fundamentally mitigates performance degradation caused by low-contrast tissue boundaries. Inspired by clinical practice, where sonographers routinely adjust image contrast to obtain clearer and more discriminative views, we incorporate this insight into the design of ACAM. By integrating contrast adjustment directly into feature learning through a dynamically parameterized module, ACAM generates anatomically meaningful multi-contrast views guided by local texture cues, significantly enhancing detail discriminability without compromising semantic extraction. Its seamless integration across convolutional, lightweight, and modern architectures demonstrates universal effectiveness, with an average accuracy gain of 1.48% validated on multi-center clinical data. Furthermore, we validated through Grad-CAM heatmaps that the proposed module enables the model to focus more on detailed information. Future work will explore physician-guided training via adjustment records and dynamic parameterization for broader contrast scenarios. ACAM provides a practical way of embedding imaging physics into deep learning pipelines, contributing to more reliable medical image analysis under heterogeneous clinical conditions.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author contributions

YC: Conceptualization, Data curation, Methodology, Resources, Visualization, Writing – original draft, Writing – review and editing. SZ: Conceptualization, Software, Writing – original draft, Writing – review and editing. BC: Writing – original draft, Writing – review and editing. MG: Conceptualization, Data curation, Writing – original draft, Writing – review and editing.

Funding

The authors declare that financial support was received for the research and/or publication of this article. This work was supported by the Fujian Province Young and Middle-aged Teachers Education Research Project (Science and Technology), Grant No. JAT200457.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphys.2025.1689936/full#supplementary-material

References

Al-Dahim A. H., Al-Anzi K. A., Al-Mutairi H. M., Al-Shammari A. M., Al-Anzi B. K., AlRashidi S. M., et al. (2024). Radiological imaging in pregnancy: balancing maternal and fetal safety. J. Int. Crisis Risk Commun. Res. 7, 1437.

Google Scholar

Bansal S., Madisetty S., Rehman M. Z. U., Raghaw C. S., Duggal G., Kumar N., et al. (2024). A comprehensive survey of mamba architectures for medical image analysis: classification, segmentation, restoration and beyond. arXiv Prepr. arXiv:2410.02362.

Google Scholar

Bock S., Weiß M. (2019). “A proof of local convergence for the adam optimizer,” in 2019 international joint conference on neural networks (IJCNN) (IEEE), 1–8.

CrossRef Full Text | Google Scholar

Boumeridja H., Ammar M., Alzubaidi M., Mahmoudi S., Benamer L. N., Agus M., et al. (2025). Enhancing fetal ultrasound image quality and anatomical plane recognition in low-resource settings using super-resolution models. Sci. Rep. 15, 8376. doi:10.1038/s41598-025-91808-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Burgos-Artizzu X. P., Coronado-Gutiérrez D., Valenzuela-Alcaraz B., Bonet-Carne E., Eixarch E., Crispi F., et al. (2020). Evaluation of deep convolutional neural networks for automatic classification of common maternal fetal ultrasound planes. Sci. Rep. 10, 10200. doi:10.1038/s41598-020-67076-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Cai J., Li Y., Liu B., Wu Z., Zhu S., Chen Q., et al. (2024). Developing deep lstms with later temporal attention for predicting covid-19 severity, clinical outcome, and antibody level by screening serological indicators over time. IEEE J. Biomed. Health Inf. 28, 4204–4215. doi:10.1109/JBHI.2024.3384333

PubMed Abstract | CrossRef Full Text | Google Scholar

Cimpoi M., Maji S., Vedaldi A. (2015). “Deep filter banks for texture recognition and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 3828–3836.

Google Scholar

Diniz P. H., Yin Y., Collins S. (2020). Deep learning strategies for ultrasound in pregnancy. Eur. Med. J. Reproductive Health 6, 73–80. doi:10.33590/emjreprohealth/20-00100

PubMed Abstract | CrossRef Full Text | Google Scholar

Fiorentino M. C., Migliorelli G., Villani F. P., Frontoni E., Moccia S. (2025a). Contrastive prototype federated learning against noisy labels in fetal standard plane detection. Int. J. Comput. Assisted Radiology Surg. 20, 1431–1439. doi:10.1007/s11548-025-03400-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Fiorentino M. C., Moccia S., Cosmo M. D., Frontoni E., Giovanola B., Tiribelli S. (2025b). Uncovering ethical biases in publicly available fetal ultrasound datasets. npj Digit. Med. 8, 355. doi:10.1038/s41746-025-01739-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao X., Wu X., Xu P., Guo S., Liao M., Wang W. (2020). Semi-supervised texture filtering with shallow to deep understanding. IEEE Trans. Image Process. 29, 7537–7548. doi:10.1109/tip.2020.3004043

CrossRef Full Text | Google Scholar

Gunasekaran S., Vivekasaran S. (2024). Disease prognosis of fetal heart’s four-chamber and blood vessels in ultrasound images using cnn incorporated vgg 16 and enhanced drnn. Int. Arab J. Inf. Technol. 21. doi:10.34028/iajit/21/6/13

CrossRef Full Text | Google Scholar

Han B., Hu M., Wang X., Ren F. (2022). A triple-structure network model based upon mobilenet v1 and multi-loss function for facial expression recognition. Symmetry 14, 2055. doi:10.3390/sym14102055

CrossRef Full Text | Google Scholar

Hou D., Cheng R., Zhang B., Wan Q., Shi P. (2025). Fault diagnosis and classifications of rolling mill bearing-gear based on gadf-tl-shufflenet-v2. J. Vib. Control, 10775463251336975. doi:10.1177/10775463251336975

CrossRef Full Text | Google Scholar

Jin Z., Zhao S., Fan S., Javdanian H. (2025). An evolutionary approach to predict slope displacement of earth embankments under earthquake ground motions. J. Eng. Res.

Google Scholar

Kashyap S., Gupta A., Ansari M. A., Singh D. K. (2023). “Review of an evolved dnn architecture efficient net for yoga pose detection problem,” in 2023 IEEE 11th region 10 humanitarian technology conference (R10-HTC) (IEEE), 829–834.

Google Scholar

Krishna T. B., Kokil P. (2023). Automated classification of common maternal fetal ultrasound planes using multi-layer perceptron with deep feature integration. Biomed. Signal Process. Control 86, 105283. doi:10.1016/j.bspc.2023.105283

CrossRef Full Text | Google Scholar

Krishna T. B., Kokil P. (2024). Standard fetal ultrasound plane classification based on stacked ensemble of deep learning models. Expert Syst. Appl. 238, 122153. doi:10.1016/j.eswa.2023.122153

CrossRef Full Text | Google Scholar

Liu X., Peng H., Zheng N., Yang Y., Hu H., Yuan Y. (2023). “Efficientvit: memory efficient vision transformer with cascaded group attention,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 14420–14430.

Google Scholar

Maher S., Seed M. (2024). Fetal cardiovascular mr imaging. Magn. Reson. Imaging Clin. 32, 479–487. doi:10.1016/j.mric.2024.04.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Mao A., Mohri M., Zhong Y. (2023). “Cross-entropy loss functions: theoretical analysis and applications,” in International conference on machine learning (pmlr), 23803–23828.

Google Scholar

Mehta K. S., Lee J. J., Taha A. A., Avgerinos E., Chaer R. A. (2017). Vascular applications of contrast-enhanced ultrasound imaging. J. Vasc. Surg. 66, 266–274. doi:10.1016/j.jvs.2016.12.133

PubMed Abstract | CrossRef Full Text | Google Scholar

Migliorelli G., Fiorentino M. C., Di Cosmo M., Villani F. P., Mancini A., Moccia S. (2024). On the use of contrastive learning for standard-plane classification in fetal ultrasound imaging. Comput. Biol. Med. 174, 108430. doi:10.1016/j.compbiomed.2024.108430

PubMed Abstract | CrossRef Full Text | Google Scholar

Miller D. L., Abo A., Abramowicz J. S., Bigelow T. A., Dalecki D., Dickman E., et al. (2020). Diagnostic ultrasound safety review for point-of-care ultrasound practitioners. J. Ultrasound Med. 39, 1069–1084. doi:10.1002/jum.15202

PubMed Abstract | CrossRef Full Text | Google Scholar

Montero A., Bonet-Carne E., Burgos-Artizzu X. P. (2021). Generative adversarial networks to improve fetal brain fine-grained plane classification. Sensors 21, 7975. doi:10.3390/s21237975

PubMed Abstract | CrossRef Full Text | Google Scholar

Mykula H., Gasser L., Lobmaier S., Schnabel J. A., Zimmer V., Bercea C. I. (2024). “Diffusion models for unsupervised anomaly detection in fetal brain ultrasound,” in International workshop on advances in simplifying medical ultrasound (Springer), 220–230.

Google Scholar

Ou Z., Bai J., Chen Z., Lu Y., Wang H., Long S., et al. (2024). Rtseg-net: a lightweight network for real-time segmentation of fetal head and pubic symphysis from intrapartum ultrasound images. Comput. Biol. Med. 175, 108501. doi:10.1016/j.compbiomed.2024.108501

PubMed Abstract | CrossRef Full Text | Google Scholar

Pei X., Huang T., Xu C. (2025). Efficientvmamba: atrous selective scan for light weight visual mamba. Proc. AAAI Conf. Artif. Intell. 39, 6443–6451. doi:10.1609/aaai.v39i6.32690

CrossRef Full Text | Google Scholar

Salman H., Parks C., Swan M., Gauch J. (2023). “Orthonets: orthogonal channel attention networks,” in 2023 IEEE international conference on big data (BigData) (IEEE), 829–837.

CrossRef Full Text | Google Scholar

Sangeetha A., Geetha P. (2024). “Survey of convnext as a cutting-edge approach in detecting polycystic ovary syndrome with advanced image analysis,” in 2024 5th international conference on data intelligence and cognitive informatics (ICDICI) (IEEE), 1391–1396.

CrossRef Full Text | Google Scholar

Smith S. W., Lopez H. (1982). A contrast-detail analysis of diagnostic ultrasound imaging. Med. Phys. 9, 4–12. doi:10.1118/1.595218

PubMed Abstract | CrossRef Full Text | Google Scholar

Szegedy C., Vanhoucke V., Ioffe S., Shlens J., Wojna Z. (2016). “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2818–2826.

Google Scholar

Venkatareddy D., Reddy K. N., Sowmya Y., Madhavi Y., Asmi S. C., Moturi S. (2024). “Explainable fetal ultrasound classification with cnn and mlp models,” in 2024 first international conference on innovations in communications, electrical and computer engineering (ICICEC) (IEEE), 1–7.

CrossRef Full Text | Google Scholar

Wang D. (2018). Image guidance technologies for interventional pain procedures: ultrasound, fluoroscopy, and ct. Curr. Pain Headache Rep. 22 (6), 6. doi:10.1007/s11916-018-0660-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang Y., Ge X., Ma H., Qi S., Zhang G., Yao Y. (2021). Deep learning in medical ultrasound image analysis: a review. Ieee Access 9, 54310–54324. doi:10.1109/access.2021.3071301

CrossRef Full Text | Google Scholar

Wittek A., Strizek B., Recker F. (2025). Innovations in ultrasound training in obstetrics. Archives Gynecol. Obstetrics 311, 871–880. doi:10.1007/s00404-024-07777-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Xu W., Fu Y.-L., Zhu D. (2023). Resnet and its application to medical image processing: research progress and challenges. Comput. Methods Programs Biomed. 240, 107660. doi:10.1016/j.cmpb.2023.107660

PubMed Abstract | CrossRef Full Text | Google Scholar

Yousefpour Shahrivar R., Karami F., Karami E. (2023). Enhancing fetal anomaly detection in ultrasonography images: a review of machine learning-based approaches. Biomimetics 8, 519. doi:10.3390/biomimetics8070519

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang L., Yang G., Ye X. (2019). Automatic skin lesion segmentation by coupling deep fully convolutional networks and shallow network with textons. J. Med. Imaging 6, 024001. doi:10.1117/1.JMI.6.2.024001

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao S., Deng H., Huang X., Xie R., Long X., Gustaf M. (2024a). Research on evaluation of expressway system operation and maintenance resilience based on dbo-elm model. Front. Phys. 13, 1647241.

Google Scholar

Zhao S., Li Z., Deng H., You X., Tong J., Yuan B., et al. (2024b). Spatial-temporal evolution characteristics and driving factors of carbon emission prediction in china-research on arima-bp neural network algorithm. Front. Environ. Sci. 12, 1497941. doi:10.3389/fenvs.2024.1497941

CrossRef Full Text | Google Scholar

Zhao S., Cao J., Lu K., Steve J. (2025). Research on olympic medal prediction based on ga-bp and logistic regression model. F1000Research 14, 245. doi:10.12688/f1000research.161865.3

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu S., Liu S., Li Y., Lei Q., Hou H., Jiang H., et al. (2024). Covidllm: a robust large language model with missing value adaptation and multi-objective learning strategy for predicting disease severity and clinical outcomes in Covid-19 patients. Curr. Proteomics 21, 591–605. doi:10.2174/0115701646366019250304064012

CrossRef Full Text | Google Scholar

Zhu S., Cai J., Xiong R., Zheng L., Chen Y., Ma D. (2025a). Contrast and gain-aware attention: a plug-and-play feature fusion attention module for torso region fetal plane identification. Ultrasound Med. Biol. 51, 2258–2266. doi:10.1016/j.ultrasmedbio.2025.08.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu S., Cai J., Xiong R., Zheng L., Ma D. (2025b). Singular pooling: a spectral pooling paradigm for second-trimester prenatal level ii ultrasound standard fetal plane identification. IEEE Trans. Circuits Syst. Video Technol., 1. doi:10.1109/tcsvt.2025.3588395

CrossRef Full Text | Google Scholar

Zhu S., Liu S., Xiong R., Zheng L., Ma D., Chen R., et al. (2025c). Multi-contrast fusion module: an attention mechanism integrating multi-contrast features for fetal torso plane classification. arXiv Prepr. arXiv:2508.09644.

Google Scholar

Keywords: fetal ultrasound, clinically-inspired module, adaptive contrast adjustment, robust medical image analysis, plug and play (PnP)

Citation: Chen Y, Zhao S, Chen B and Gustaf M (2025) Clinically guided adaptive contrast adjustment for fetal plane classification: a modular plug-and-play solution. Front. Physiol. 16:1689936. doi: 10.3389/fphys.2025.1689936

Received: 22 August 2025; Accepted: 03 November 2025;
Published: 13 November 2025.

Edited by:

Choon Hwai Yap, Imperial College London, United Kingdom

Reviewed by:

Maria Chiara Fiorentino, Marche Polytechnic University, Italy
Vinayakumar Ravi, Prince Mohammad bin Fahd University, Saudi Arabia

Copyright © 2025 Chen, Zhao, Chen and Gustaf. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Måns Gustaf, Z3VzdGFmZWR1QHllYWgubmV0

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.