ORIGINAL RESEARCH article

Front. Physiol., 28 April 2025

Sec. Computational Physiology and Medicine

Volume 16 - 2025 | https://doi.org/10.3389/fphys.2025.1559389

This article is part of the Research TopicMedical Knowledge-Assisted Machine Learning Technologies in Individualized Medicine Volume IIView all 14 articles

Deep learning-based automated tongue analysis system for assisted Chinese medicine diagnosis

Tingnan ChenTingnan Chen1Yutong ChenYutong Chen1Zili ZhouZili Zhou2Ying ZhuYing Zhu1Ling HeLing He1Jing Zhang
Jing Zhang1*
  • 1College of Biomedical Engineering, Sichuan University, Chengdu, China
  • 2Sichuan Second Hospital of TCM, Chengdu, China

This study proposes an automated tongue analysis system that combines deep learning with traditional Chinese medicine to enhance the accuracy and objectivity of tongue diagnosis. The system includes a hardware device to provide a stable acquisition environment, an improved semi-supervised learning segmentation algorithm based on U2net, a high-performance colour correction module for standardising the segmented images, and a tongue image analysis algorithm that fuses different features according to the characteristics of each feature of the TCM tongue image. Experimental results demonstrate the system’s performance and robustness in feature extraction and classification. The proposed methods ensure consistency and reliability in tongue analysis, addressing key challenges in traditional practices and providing a foundation for future correlation studies with endoscopic findings.

1 Introduction

The diagnosis of the tongue in traditional Chinese medicine is a method used to assess the health of internal organs by observing characteristics such as shape, color, texture, and coating of the tongue. However, traditional tongue diagnosis relies heavily on the experience of doctors, making it susceptible to subjective biases. With advancements in artificial intelligence and medical imaging technologies, intelligent tongue diagnosis systems based on image processing have emerged, providing an objective and standardized approach to tongue analysis. Nevertheless, due to variations in the photographic environment and lighting conditions, the tongue images of the same patient can differ significantly. Therefore, ensuring standardization in image acquisition and achieving precise image processing are key research challenges in the field of tongue diagnosis.

This paper proposes a deep learning-based tongue analysis system that uses image processing techniques to achieve automated segmentation and analysis of tongue images, ensuring stability and consistency in results. Specifically, we propose a semi-supervised learning-based tongue segmentation algorithm built on the U2Net model, incorporating several innovative modules for precise feature extraction and evaluation. The key innovations in this paper are as follows: (1) Development of a stable hardware and software environment for tongue image acquisition, ensuring standardized and rational image assessment. (2) An improved semi-supervised method based on the U2Net model, which effectively captures the scale characteristics of tongue segmentation and enhances model performance and generalization by utilizing a large number of unlabeled images. (3) A precise color correction method for the tongue image acquisition device, facilitating more standardized and accurate classification in subsequent analysis. (4) A comprehensive tongue diagnosis framework based on the image characteristics of different tongue types, achieving excellent performance even in challenging classification scenarios.

2 Related work

Tongue diagnosis has a long history in traditional Chinese medicine. With advancements in technology, digital image processing methods for tongue diagnosis have gradually developed. In recent years, many researchers have attempted to apply tongue image segmentation, color correction, and feature extraction in automated tongue diagnosis systems.

Existing tongue segmentation methods primarily include traditional image processing techniques and deep learning models. Early studies utilized edge detection and color thresholding methods for tongue segmentation; however, these approaches performed poorly in complex backgrounds. With the development of deep learning, convolutional neural networks (CNNs) have become the main tools for tongue segmentation. Kang et al. (2024) cascaded YOLOv5 with the LA-Unet network to refine the segmentation of tongue regions, optimizing segmentation for mobile tongue images. Zhang et al. (2022b) performed structural optimization on the DeeplabV3+ network, leveraging prior knowledge of tongue images to enhance edge regions, achieving precise results. Gao et al. (2021) proposed a level set model with symmetry and edge constraints, combining geometric features of the tongue for segmentation, capable of handling tongue images in most conditions. However, these segmentation methods only generalized their training to different imaging devices, making it challenging to achieve precise classification and recognition of tongue images, while not fully utilizing unlabeled tongue images. In other areas of biomedicine, Zhong et al. (2022) unsupervised approach to deconvolution in genomic subclones. Zhang et al. (2022a) proposing a lesion-aware dynamic network (LDNet) for polyp segmentation, which is a conventional U-shaped encoder-decoder structure combined with a dynamic kernel generation and update scheme. Gao et al. (2022) proposed a novel weak semi-supervised framework called SOUSA (Segmentation Using Only Sparse Annotations), which aims to learn from a small number of sparsely annotated datasets and a large amount of unlabeled data. Zhao et al. (2022) propose a cross-level contrast learning scheme to enhance the representation of local features in semi-supervised medical image segmentation. Inspired by these previous studies and given the scarcity of specialized tongue images, this paper proposes an improved U2Net tongue segmentation model combined with semi-supervised learning, enabling high-precision segmentation of tongue images captured by professional equipment, with strong interference resistance and robustness against abnormal tongue conditions. This approach effectively addresses the fragmentation of existing tongue analysis methods, enhancing their practical application value.

Furthermore, in terms of color correction for tongue images, existing research typically employs color mapping methods to address color discrepancies caused by different shooting devices and lighting conditions. Sun et al. Xin et al. (2021) proposed a gray world-based rapid color correction method for tongue images, assessing the degree of distortion after image compression, followed by color correction based on this degree, improving the effectiveness of color correction. Yan et al. (2021) introduced a TCCGAN network to correct tongue image colors, initially employing a differentiable weighted histogram network for color feature extraction, utilizing a new upsampling module called mixed feature attention upsampling to assist in image generation, while constructing a stacked network to generate tongue images from coarse to fine. However, enhancing the generalization ability of color correction methods to adapt to complex clinical environments remains a challenge. This paper proposes a space-distance-weighted Lasso regression algorithm, optimized for the regression environment of each color, effectively addressing issues such as image distortion and color overfitting after correction, laying a solid foundation for subsequent analysis of tongue and coating colors.

In tongue classification tasks, researchers have traditionally relied on texture feature extraction and machine learning classifiers for tongue evaluation. Recently, with the widespread application of deep learning models, significant advancements have been made in neural network-based tongue feature extraction and classification models. Chen et al. (2023) employed the K-means algorithm for coating separation and utilized RGB components of images to assess tongue color, achieving a balance between simplicity and accuracy. Yiqin (2022) used a Gaussian mixture model to separate tongue coating from the tongue body, developing a model for tongue image restoration, and ultimately achieved good results in classifying tongue textures based on ResNet101. Zhang et al. (2023) implemented a dual-threshold segmentation method based on HSI color space to automatically extract tongue bodies from original tongue images, categorizing tongues into those with and without coating, and further classifying coating thickness based on area. In the biomedical field, there are also researchers who focus on image detail features using target detection methods, Zhao et al. Prisilla et al. (2023) identified which YOLO models (YOLOv5, YOLOv6, and YOLOv7) performed well in detecting LDH in different regions of the lumbar disc. However, the complex boundaries and diverse texture information in tongue images continue to pose challenges for models. Thus, designing models that address the complex features of tongue coatings has become a current research focus. This paper presents a tongue feature judgment module based on different tongue feature groupings and precise coating separation, employing a precisely annotated coating separation dataset and a high-performance GSCNN model capable of handling complex boundaries, while utilizing LBP images and wavelet fusion features to significantly improve accuracy in difficult tongue classification.

Additionally, some researchers have combined other medical features with machine learning to implement end-to-end tongue applications. Yuan et al. (2023) integrated tongue coating microbiomes to establish an AI deep learning model, evaluating the value of tongue images and microbiomes in gastric cancer diagnosis. In other areas of biomedicine, Syed (2023) develop an automatic deep learning-based brain atrophy diagnosis model to detect, segment, classify, and predict the survival rate. Alabi et al. (2022) discussed deep learning technical knowledge and algorithms for OSCC and the application of deep learning techniques to cancer detection, image classification, segmentation and synthesis, and treatment planning. Zhou et al. (2022) summarized the workflows of deep learning methods in medical images and the current applications of deep learning-based AI for diagnosis and prognosis prediction in bone tumors. Having received help in working with the above systems, this paper proposes a comprehensive tongue diagnosis system that integrates tongue image acquisition, segmentation, color correction, and judgment, allowing accurate and efficient capture of key features of patients’ tongue images.

3 Clinical samples

The clinical samples in this study were obtained from 2,738 patients at Sichuan Province Second Hospital of Traditional Chinese Medicine, Mianzhu City Traditional Chinese Medicine Hospital, Guanghan City Traditional Chinese Medicine Hospital, and Anyue County People’s Hospital. Tongue images of these patients were collected using a specially designed tongue imaging device, and they underwent gastrointestinal endoscopy. Ethics approval for this study was obtained from Medical Ethics Committee of Sichuan Second Hospital of Traditional Chinese Medicine [approval number: 202304(H)-003-01]. All patients signed informed consent forms, agreeing to the use of their clinical samples for this research, and provided personal information including name, age, ethnicity, occupation, medical history and use of medications.

The tongue image data for each patient, including tongue body area, tongue coating area, tongue color, tongue texture (including tooth-marked tongue and cracked tongue), tongue shape, coating color, and coating texture, were annotated by professional physicians from the aforementioned hospitals. Due to the complex boundaries of the tongue coating area, pixel-level annotation was challenging for regular annotation tools; therefore, high-standard pixel-level annotations were conducted using Photoshop.

The tongue color was categorized into five types: pale, light red, red, dark red, and others. The coating color was classified into white, yellow, and other types. Additionally, various tongue conditions were classified as follows: tooth-marked tongue (presence or absence), cracked tongue (presence or absence), thickness (thick or thin coating), spotted tongue (presence or absence), peeling coating (presence or absence), curdy or slimy coating (curdy coating, slimy coating, or normal coating), and moist-dry condition (slippery coating, moist coating, or dry coating).

4 Materials and methods

A complete tongue image segmentation and evaluation system was established. As shown in Figure 1, the system consists of four modules: tongue image acquisition hardware module, segmentation module, color correction module, and evaluation module.

Figure 1
www.frontiersin.org

Figure 1. Flowchart of the intelligent tongue diagnosis system.

4.1 Tongue image acquisition hardware module

As shown in Figure 2,The acquisition box provides a constant light source to stabilise the tongue image acquisition environment, a customised industrial camera is used to provide high quality images, and the box is designed with a tilted angle to facilitate the presentation of the patient’s tongue. A mandibular rest is also used for the purpose of positioning the tongue.

Figure 2
www.frontiersin.org

Figure 2. Appearance of the structure of the tongue filming equipment.

4.2 Tongue image segmentation module

A semi-supervised learning algorithm based on the U2Net model is proposed to address the characteristics of tongue images. In the acquired tongue images, most of the pixels are occupied by the subject’s face and tongue, and the segmentation involves only the foreground and background, making it a salient object detection (SOD) task. Previous studies have used conventional segmentation networks like Mask R-CNN and DeepLabv3+ for tongue segmentation. Although these networks achieve excellent segmentation performance, their pre-trained backbone networks are typically trained on ImageNet, which limits their performance on specific tongue image segmentation tasks. To overcome these issues, we used U2Net as the main segmentation model structure to achieve stability in this SOD task for tongue segmentation. Since pixel-level labeling of tongue images requires substantial time and effort, we adapted it to a semi-supervised model to make full use of the data and achieve optimal training results.

U2Net (U-square-Net) Qin et al. (2020) is a deep learning-based segmentation model particularly suitable for salient object detection and segmentation tasks. Its name comes from its unique nested U-shaped structure. This model improves upon the traditional U-Net by incorporating several smaller U-Nets (called Residual U-blocks or RSU modules) to enhance feature extraction and multi-scale feature aggregation capabilities. The overall model consists of a U-shaped structure with 11 stages, each containing an RSU module, forming a six-level encoder and a five-level decoder, with skip connections between corresponding encoder and decoder layers to fuse multi-scale features.

To better suit the characteristics of tongue images, we adapted U2Net into a semi-supervised model called U2Net-MT. Mean Teacher Tarvainen and Valpola (2017) is a deep learning model for semi-supervised learning, in which a teacher network and a student network are used. The parameters of the teacher network serve as the targets for the student network, allowing the student to gradually learn more accurate parameters. This method effectively utilizes unlabeled data and enhances model generalization, especially in scenarios with limited labeled data.

As shown in Figure 3, during training, both labeled tongue image A and unlabeled tongue image B are input into both the U2Net student model and the teacher model (which share the same structure). Random noise and image augmentations (including horizontal flipping and random cropping) are added to the different inputs in each model.

Figure 3
www.frontiersin.org

Figure 3. U2Net-MT Network. (a) U2Net-MT Network Architecture. (b) Schematic Diagram of U2Net Network Architecture Qin et al. (2020).

4.2.1 Encoder and decoder

In the U2Net model, the input image I is processed by the encoder to generate a set of hierarchical features M={m1,m2,,m6}. For each feature mi, the decoder processes it hierarchically. In the decoder, each level’s input is obtained by concatenating (using the Concat operation) the output of the lower-level decoder and the corresponding encoder output, expressed as Equation 1:

di=Concatui+1,mi(1)

where ui+1 is the output of the lower-level decoder, and mi is the output of the corresponding encoder. Each scale’s output Oi in the decoder undergoes a 3×3 convolution followed by a sigmoid activation to generate the saliency probability map Si, given by Equation 2:

Oi=σWidi+bi(2)

where Wi represents the convolution kernel, denotes the convolution operation, bi is the bias term, and σ is the sigmoid activation function defined as Equation 3:

σx=11+ex(3)

The output Oi is upsampled to match the input image size, and then concatenated with the input image. After a 1×1 convolution and sigmoid activation, the final saliency probability map Sfuse is obtained, given by Equations 4, 5:

Oiup=UpsampleOi(4)
Sfuse=σWfuseConcatOiup,I+bfuse(5)

4.2.2 Loss function

For the labeled tongue image A, the loss L1 is calculated for the fused probability map Sa using the binary cross-entropy (BCE) loss (Equation 6):

L1=1Ni=1NyilogSai+1yilog1Sai(6)

For all input images, saliency maps Sis and Sit are obtained for each level’s decoder from the student and teacher models, respectively. A loss function L2i is calculated at each level using the mean squared error (MSE) between Sis and Sit (Equation 7):

L2i=1Nj=1NSisjSitj2(7)

A confidence weighting module is introduced to strengthen the loss function. The total consistency loss L2 is computed by weighting each L2i with the average pixel confidence Confi of Sis and Sit (Equation 8):

L2=iωiL2iiωi(8)

where ωi is calculated as the average confidence of Sis and Sit. The overall loss of the model is given by (Equation 9):

L=L1+λL2(9)

where λ is the weighting coefficient for the loss. During training, L is minimized to obtain the final model parameters.

The trained student model is used for inference to evaluate its performance.

4.3 Color correction module

Equations should be inserted in editable format from the equation editor. In this experiment, all tongue images were captured using a mature, fixed hardware system. Therefore, all images undergo a strict color correction process to ensure data accuracy and consistency. During the experiment, ColorChecker 24 color card images were taken using different hardware devices of the same model, and various regression methods were used to calibrate the RGB values of the images to match the true values. Below, the basic theory of color correction and the applied color correction module are described.

4.3.1 Basic principles of color correction

In color correction, the aim is to find the mapping relationship between the RGB values of the image and the standard color values of the ColorChecker 24 color card. Instead of performing regression on each color channel (R, G, B) separately, all RGB values are treated as a whole for regression analysis, which can account for correlations among RGB values and achieve more accurate color correction results.

4.3.1.1 Color correction process

A multiple regression model is built, including RGB values, to directly map the original RGB values to the corrected RGB values. The regression model can be represented as Equation 10:

RcorrectedGcorrectedBcorrected=XβRβGβB(10)

where X is the design matrix containing the original RGB values, represented as RGB1. βR,βG,βB are the regression coefficients to be solved, which map the original RGB values to corrected RGB values. Rcorrected,Gcorrected,Bcorrected represent the corrected red, green, and blue channel values.

In matrix form, this can be expressed as Equation 11:

Y=Xβ(11)

where Y contains the corrected RGB values, represented as RcorrectedGcorrectedBcorrected, and β is the regression coefficient matrix (Equation 12):

βR1βR2βR3βR0βG1βG2βG3βG0βB1βB2βB3βB0(12)

The regression coefficients are solved using the least squares method. Given n ColorChecker 24 samples and their standard values, the optimization problem can be expressed as Equation 13:

minβi=1nYtrue,iXiβ2(13)

where Ytrue,i is the true RGB value of the i-th sample, and Xi is the original RGB value matrix of the i-th sample.

The best regression coefficients β can be solved using the matrix Equation 14:

β=XTX1XTYtrue(14)

Using the obtained regression coefficients β, each pixel of a new image can be corrected. The steps are as follows:

For each pixel (R,G,B), form an input vector (Equation 15):

X=RGB1(15)

Use the regression model to calculate the corrected RGB values (Equation 16):

Ycorrected=Xβ(16)

Obtain the corrected RGB values (Rcorrected,Gcorrected,Bcorrected).

4.3.2 Lasso regression algorithm based on spatial distance weighting

ColorChecker 24 includes 24 color patches, covering a limited range of the color spectrum. This limitation may lead to overfitting issues, and the mapping derived from fitting these 24 colors often lacks generalization, making it less effective for correcting colors outside the range of the color patches.

To address this issue, this work employs a lasso regression algorithm based on spatial distance weighting.

4.3.2.1 Calculating the distance and weight from pixels to ColorChecker 24

Suppose the RGB values of ColorChecker 24 in the XYZ Cartesian coordinate system are labeled as CCn(Rn,Gn,Bn), where n=1,2,,24.

As shown in Figure 4 for each pixel P(R,G,B) in the image, calculate its Euclidean distance to each CCn (Equation 17):

Ln=RRn2+GGn2+BBn2(17)

Figure 4
www.frontiersin.org

Figure 4. Spatial distribution of ColorChecker 24 patches and Euclidean distances to the target pixel.

Based on the calculated distance Ln, determine the regression weight of each ColorChecker point CCn for point P (Equation 18):

wn=1Ln+ω(18)

where ω is a dilution term to prevent excessively large weights when Ln is small.

4.3.2.2 Polynomial feature transformation and lasso regression

Assume the degree of polynomial transformation is D. For each CCn and P, the transformed feature vector is Equation 19:

PolyCCn=1,Rn,Gn,Bn,Rn2,RnGn,RnBn,Gn2,GnBn,Bn2,,RnD,GnD,BnDPolyP=1,R,G,B,R2,RG,RB,G2,GB,B2,,RD,GD,BD(19)

The weighted lasso regression model is then used to fit the data (Equation 20):

FP=Lassoα=0.01P(20)

The loss function of the model is (Equation 21):

12mi=1mwiyiyî2+αj=1n|θj|(21)

where wi is the sample weight, yi is the actual value, yî is the predicted value, α is the regularization parameter, and θj are the regression coefficients.

Use the trained lasso model to predict the corrected value for each pixel, and combine the corrected pixel values into the corrected image.

4.4 Tongue image analysis module

The tongue image analysis module consists of the tongue coating color determination module and the coating texture determination module. As shown in Figure 5, the color-corrected segmented tongue image is first input into the tongue coating color determination module. Here, the coating texture is separated into tongue coating and tongue body images, and the respective tongue and coating colors are determined. Finally, the color-corrected segmented tongue image is input into the coating texture determination module to obtain the texture analysis results.

Figure 5
www.frontiersin.org

Figure 5. The network structure of the tongue image analysis module.

4.4.1 Color determination module

A deep learning model is used to separate the segmented tongue image into the coating and body images. Due to the complex boundaries and pixel-level distribution of tongue coatings, Gated-SCNN (Gated Shape Convolutional Neural Network) is introduced for coating separation.

Gated-SCNN (Gated Shape CNN) Takikawa et al. (2019) is an improved convolutional neural network architecture specifically designed for image segmentation tasks. Its main innovation is the introduction of a shape stream and gating mechanism. Structurally, Gated-SCNN consists of two parallel branches: the backbone network, responsible for extracting semantic features from the image, typically using ResNet or VGG; and the shape stream, which focuses on capturing edges and contours through a shape convolution module to extract multi-scale shape features. The shape information is fused with the semantic features using the gating mechanism, which dynamically adjusts the weight of the shape information at different locations. This combination enhances the ability of the model to handle complex boundaries and details, significantly improving segmentation accuracy.

After applying GSCNN to the input image, the result is inverted with the non-black regions of the original image to obtain the tongue body image.

For color determination of the tongue body and coating, color moments are used as feature extraction methods. Color moments Stricker and Orengo (1995) include the mean, variance, and skewness of color components. For each color component (e.g., red, green, and blue), the following three statistics are extracted:

1. Mean 2. Variance 3. Skewness.

These features are arranged in a specific order to form a nine-dimensional vector representing the color characteristics. Specifically, given a color component C, its mean, variance, and skewness are calculated as follows (Equations 2224):

μC=1Ni=1NCi(22)
σC2=1Ni=1NCiμC2(23)
γC=1Ni=1NCiμCσC3(24)

Where Ci represents the color value of the i-th pixel, N is the total number of pixels, μC is the mean, σC2 is the variance, and γC is the skewness of the color component.

The calculated mean, variance, and skewness are arranged in the order of red, green, and blue components to form a nine-dimensional feature vector (Equation 25):

F=μR,σR2,γR,μG,σG2,γG,μB,σB2,γB(25)

This nine-dimensional vector F is then fed into a trained classifier to determine the tongue and coating color.

4.4.2 Tongue image classification module

In the tongue image classification module, the features are divided into three categories:

Category 1: Teeth marks, cracks, thickness, and spots.

Category 2: Peeling and curdy or slimy.

Category 3: Moistness and dryness.

For Category 1, the features are clear and the deep learning network can easily extract them for direct classification. For Category 2, where peeling and curdy or slimy coatings are to be differentiated from non-peeling and normal coatings, the grayscale images are analyzed using Local Binary Pattern (LBP) operator for feature extraction.

The Local Binary Pattern (LBP) operator Ojala et al. (1994) is used for texture feature extraction. Given an input image I with a pixel at (x,y):

The LBP value at each pixel is calculated as follows (Equation 26):

LBPx,y=p=0P1sIxp,ypIx,y2p(26)

Where (xp,yp) are the neighboring pixels of (x,y), P is the number of neighbors, and s(x) is a sign function defined as Equation 27:

sx=1if x00if x<0(27)

The LBP values for each pixel are stored at the corresponding location in the output image ILBP (Equation 28):

ILBPx,y=LBPx,y(28)

Where ILBP is of the same size as the original image I.

For Category 3, which deals with subtle texture changes, neural networks alone are insufficient. It was found experimentally that a combination of wavelet transformation and deep learning improved performance for distinguishing between moist and dry coatings.

In this experiment, the Daubechies wavelet (db3) was selected as the wavelet basis function. After the wavelet transformation, three high-frequency detail images are obtained: horizontal details (CH), vertical details (CV), and diagonal details (CD). To ensure the resolution matches that of the original image (224×224 pixels), interpolation was applied to these detail images.

The original tongue image IRH×W×C, where H and W represent the image height and width respectively, and C is the number of color channels. After wavelet transformation (Equation 29):

ICH,ICV,ICD=DWTI,db3(29)

Where DWT represents discrete wavelet transform and db3 is the wavelet basis.

Each detail image is resized back to 224×224 resolution (Equations 3032):

ICH=InterpICH,224,224(30)
ICV=InterpICV,224,224(31)
ICD=InterpICD,224,224(32)

Where Interp represents interpolation.

Next, the processed high-frequency detail images ICH, ICV, and ICD are concatenated with the original color channels IR, IG, and IB along the channel dimension, resulting in a concatenated image with six channels Istacked (Equation 33):

Istacked=IR,IG,IB,ICH,ICV,ICD(33)

To achieve feature fusion, Depthwise Separable Convolution Howard et al. (2017) was used. Depthwise convolution applies a 3×3 convolution kernel to each input channel independently, as shown by Equation 34:

Yd=IstackedKd(34)

where Yd represents the output after depthwise convolution, Kd is the depthwise convolution kernel, and denotes the convolution operation.

Subsequently, pointwise convolution uses a 1×1 convolution kernel to combine information across all input channels (Equation 35):

Yp=YdKp(35)

where Yp is the output after pointwise convolution, and Kp is the pointwise convolution kernel.

After the convolution operations, a ReLU activation function is applied (Equation 36):

Z=ReLUYp(36)

To further enhance feature representation, the Convolutional Block Attention Module (CBAM) Woo et al. (2018) was integrated. CBAM first performs adaptive max pooling and adaptive average pooling on the input feature map to aggregate global information along the channel dimension, producing two descriptors (Equations 37, 38):

MaxPoolZ=MaxZ,dim=H×W(37)
AvgPoolZ=AvgZ,dim=H×W(38)

These descriptors are then fed into shared convolutional layers consisting of two 1×1 convolution layers. The first convolutional layer reduces the number of channels to half of the original size and uses a ReLU activation function (Equation 39):

F1=ReLUW1MaxPoolZ,AvgPoolZ(39)

The second convolutional layer restores the original channel size (Equation 40):

F2=σW2F1(40)

where σ represents the Sigmoid activation function. The output of the convolutional layers serves as channel attention weights, which are combined with the original feature map through element-wise multiplication to enhance important features.

The enhanced feature map is then processed by the spatial attention module. This module performs global max pooling and global average pooling along the spatial dimension to generate two single-channel feature maps (Equations 41, 42):

MaxPoolsZ=MaxZ,dim=C(41)
AvgPoolsZ=AvgZ,dim=C(42)

These feature maps are concatenated along the channel dimension to form a two-channel feature map, which is then processed by a 7×7 convolution layer to generate the spatial attention map (Equation 43):

Fs=σWsMaxPoolsZ,AvgPoolsZ(43)

The spatial attention map is combined with the channel-enhanced feature map through element-wise multiplication to further improve feature representation.

For the image classification task, Vision Transformer (ViT) Dosovitskiy et al. (2021) was used as the classification head. ViT divides the input image into fixed-size patches, flattens each patch, and embeds it into a high-dimensional vector space. These embedding vectors are then added to positional encodings to retain positional information and processed through multiple Transformer encoder layers.

The input image ZRH×W×C is divided into N patches of size P×P, where N=HWP2. Each patch ziRP×P×C is flattened and mapped to a high-dimensional vector through a linear transformation (Equation 44):

ei=FlattenziWe+be(44)

where WeR(P2C)×D is the embedding matrix, beRD is the bias vector, and D is the embedding dimension.

The positional encoding vector EposRN×D is added to the embedding vectors to retain positional information (Equation 45):

E0=e1;e2;;eN+Epos(45)

where E0RN×D is the initial embedding representation.

The initial embedding representation E0 is processed through L Transformer encoder layers. Each encoder layer includes a Multi-Head Self-Attention (MHSA) mechanism and a Feed-Forward Neural Network (FFN):

Multi-Head Self-Attention (Equation 46):

MHSAE=head1;head2;;headhWO(46)

where each attention head is defined as (Equation 47):

headi=AttentionEWiQ,EWiK,EWiV(47)

The attention calculation is (Equation 48):

AttentionQ,K,V=softmaxQKTdkV(48)

where Q,K,V are the query, key, and value matrices, respectively, and dk is the dimension of the keys.

Feed-Forward Neural Network (Equation 49):

FFNE=ReLUEW1+b1W2+b2(49)

where W1,W2 are weight matrices and b1,b2 are bias vectors.

The output of each Transformer encoder layer is represented as (Equations 50, 51):

El+1=LayerNormEl+MHSAEl(50)
El+2=LayerNormEl+1+FFNEl+1(51)

where LayerNorm represents the layer normalization operation.

After L Transformer encoder layers, the final feature representation EL is fed into the classification head. A class token is introduced in the feature representation, which is processed by the Transformer encoder layers and used for the final classification task (Equation 52):

class token=EL0(52)

The classification is performed using a linear transformation followed by a softmax function (Equation 53):

y=softmaxWclsclass token(53)

where Wcls is the weight matrix of the classification head.

Focal Loss was used as the loss function to handle the imbalance in tongue image features, calculated as (Equation 54):

FLpt=αt1ptγlogpt(54)

where pt represents the predicted probability, and αt and γ are hyperparameters.

5 Experiments

5.1 Evaluation metrics

For the segmentation task, Mean Absolute Error (MAE) and the Dice coefficient were used as evaluation metrics. MAE provides an intuitive measure of the overall pixel-wise classification accuracy, while the Dice coefficient considers both precision and recall for all instances, focusing on the overlap with the target regions.

For the classification task, accuracy (acc) and macro-F1 score were used. This not only considers overall accuracy but also gives more weight to frequently occurring samples.

To evaluate the performance of color correction, ΔEab* was used to quantify color differences. ΔEab* is an index for quantifying and describing color differences, based on the CIELAB color space, and it compares the visual difference between two colors. A larger value of ΔEab* indicates a more noticeable difference between the two colors. The calculation of ΔEab* is as follows (Equation 55):

ΔEab*=ΔL*2+Δa*2+Δb*2(55)

where:

ΔL* represents the difference in lightness (L*) between the two colors;

Δa* represents the difference in the red-green axis (a*) between the two colors;

Δb* represents the difference in the yellow-blue axis (b*) between the two colors.

5.2 Experimental results

5.2.1 Performance analysis

To verify the performance of the proposed method, comparisons were made with existing methods used in tongue image segmentation and classification.

UNet Ronneberger et al. (2015) is a classic image segmentation network consisting of an encoder and a decoder, which effectively extracts and restores image details through its symmetric structure and skip connections. The encoder extracts features, the decoder restores resolution, and the skip connections transmit features between the encoder and decoder, preventing feature loss and improving segmentation accuracy. UNet has been widely used in medical image processing and biological image analysis. UNet++ Zhou et al. (2020) is an improved version of UNet, adding more dense skip connections and decoder submodules. It introduces additional convolution layers at each downsampling and upsampling stage to form dense connectivity paths, capturing multi-scale features and enhancing segmentation accuracy and robustness, which is suitable for segmenting complex image structures. FCN Long et al. (2014) (Fully Convolutional Network) removes the fully connected layers and only uses convolution and upsampling layers. It extracts features through a series of convolution layers and restores the original resolution through deconvolution. FCN’s design preserves spatial information and enables efficient pixel-level classification, suitable for end-to-end segmentation of images of any size. DeepLabv3+ Chen et al. (2018) combines the encoder-decoder structure and atrous convolution, using ResNet or Xception as the backbone. The encoder extracts multi-scale features, and the decoder enhances segmentation with atrous convolution to capture more contextual information. The multi-scale feature fusion strategy improves segmentation accuracy and detail retention, suitable for image segmentation in complex scenes. Mask R-CNN Massa and Girshick (2018) adds a branch for generating pixel-level segmentation masks on top of Faster R-CNN. The backbone extracts features, the region proposal network generates candidate regions, and each candidate is classified, regressed, and mask-generated. Mask R-CNN performs object detection and instance segmentation simultaneously, providing more precise segmentation results and being widely used in instance segmentation tasks. In this experiment, semi-supervised models used a 1:1 supervision rate, and GSCNN adopted VGG16 as the backbone, with all models loaded with pre-trained weights.

The Table 1 shows the comparison of U2net-MT and GSCNN with other segmentation networks on the dataset used in this study. Meanwhile Table 2 shows a more intuitive image of the segmentation results under different models. The results indicate that both networks achieved the best performance on two metrics, demonstrating that the proposed and applied methods effectively address the segmentation and coating separation problems of the tongue dataset. Although other models listed in the table have shown excellent performance in medical image processing and semantic segmentation, they lack focused analysis of features from specific depths during segmentation, and they struggle with ambiguous boundaries between the tongue, throat, and lips. For coating separation, traditional segmentation networks face challenges in extracting effective features from the highly detailed, dispersed, and weakly correlated coating on the tongue. U2net-MT combines U2net’s strengths in retaining full-resolution features and multi-scale feature fusion while optimizing significant target detection, with a semi-supervised approach to improve data utilization and generalization capability. By assigning more weight to high-confidence scales, the model enhances scale-specific attention during semi-supervised training, effectively improving the classification performance of fuzzy pixels around the tongue. GSCNN introduces the Canny operator in features, using a gating mechanism to ensure that only boundary-related information is processed in the shape stream, enabling it to effectively handle complex coating boundaries.

Table 1
www.frontiersin.org

Table 1. Comparison of the effect of the model used in this experiment with other image segmentation model methods.

Table 2
www.frontiersin.org

Table 2. Comparison of a complete tongue image and segmentation results in different models.

The Table 3 shows the results of the proposed spatial-distance-weighted Lasso regression algorithm and other common algorithms in color correction of the ColorChecker 24 color card, measured by ΔEab*. The results show that the spatial-distance-weighted Lasso regression achieved the best results for both average and maximum/minimum ΔEab*. This is because the spatial-distance-weighted Lasso regression focuses more on the regression relationships of similar colors instead of treating the 24 colors as a whole. Regularization via Lasso regression also effectively solves overfitting in small sample regressions. In contrast, traditional linear regression methods have poor accuracy in color correction, and higher-order regression, while adapting to the specific regression characteristics of each color, can have steep gradients at color boundaries, causing visual artifacts in the corrected images. These results indicate that spatial-distance-weighted Lasso regression effectively addresses subtle lighting variations caused by different environments, providing a basis for subsequent classification of tongue and coating colors.

Table 3
www.frontiersin.org

Table 3. Comparison of ΔE*ab between the tongue colour correction method used in this paper and other commonly used colour correction methods.

To ensure the classification results of tongue diagnosis are of the highest quality, different classifiers were applied to the Color and Tongue feature sections to observe the experimental outcomes. Due to space limitations in the table, only some of the more effective methods are selected for comparison. In the Color section, the classifiers primarily focus on the processed visual color features. As a result, simpler machine learning classification methods are more effective in the experiments. Among these methods, Random Forest (RF) demonstrates excellent anti-overfitting capabilities and can also filter the importance of features, thereby validating whether the extracted features contribute significantly to classification. Regarding K-Nearest Neighbors (KNN), the color-classification boundaries in these color matrix samples should be relatively clear, making KNN quite effective. Similarly, in the more complex color matrix vectors, Support Vector Machines (SVM) can maximize the margin between classes, and its advantage over KNN is that it can be trained in advance. The Softmax method, combined with a simple neural network structure, performs excellently in multi-class classification tasks and can also be optimized through backpropagation.

In the Tongue feature section, the classical ResNet structure is also used for comparison. The reason for choosing it is its ability to extract deep features from images while being highly versatile. EfficientNet enhances computational efficiency by optimizing the structure, offering good advantages in texture feature computation. MobileNet maintains high efficiency in lightweight design, and the reason for selecting it is to observe the performance of lightweight models in tongue feature extraction and computation. Vision Transformers (ViT), relying on the self-attention mechanism and Transformer architecture, are able to capture complex patterns and details in images, and they perform well on large-scale datasets. By combining the convolutional feature extraction capability of ResNet50 with the global self-attention mechanism of ViT, both local and global information can be utilized. If these features are useful, the classification performance will be enhanced. Among these methods, the focus of this paper is on those that demonstrate efficient and stable performance across various features.

The Table 4 presents the results of the proposed tongue diagnosis algorithm with different classifiers. The extracted tongue features allowed the classifiers to effectively classify based on tongue characteristics, indicating that tongue diagnosis, as an integrated module, could effectively extract overall tongue information of patients. Selecting the appropriate classifier for different information yielded better results. In experiments, the ViT model maintained stable and excellent performance in multiple classification tasks. This is because the relatively simple structure of the transformer and the efficient attention mechanism can focus on different texture classification features. Therefore, the ViT-b model was chosen for inference in the tongue diagnosis module. Additionally, the ViT with ResNet as the backbone did not perform as well in most classification tasks, indicating that deep feature extraction is limited for guiding classification. The performance of lightweight networks was also not remarkable, reaffirming the suitability of the transformer structure for texture features of the tongue.

Table 4
www.frontiersin.org

Table 4. Performance of multiple models and classifiers in tongue image feature classification (%).

5.2.2 Ablation study

This section presents ablation experiments to validate the effectiveness of various components and key methods in each module of the tongue image system on the test set. “Without” indicates the absence of the specific key method from the module. The experiments are divided as follows: (1) Evaluation of the MT semi-supervised module and confidence-weighted module in U2net-MT on the noisy test set. (2) Evaluation of the effectiveness of the LBP features from the second group of tongue images, wavelet features from the third group, feature fusion, and the CBAM module in the tongue diagnosis module.

As shown in the Table 5, U2net-MT achieved optimal results for both metrics. Removing the MT semi-supervised method led to a significant decrease in the Dice coefficient, indicating that the MT semi-supervised method effectively enhances the model’s generalization capability, reduces the impact of noise, and improves accuracy by leveraging features from unlabeled data. Removing the confidence-weighted module resulted in a performance drop in both metrics, suggesting that the confidence-weighted module helps the model focus on feature information that is beneficial for tongue image segmentation during backpropagation. These findings demonstrate the effectiveness of the methods used in U2net-MT for training in tongue image segmentation.

Table 5
www.frontiersin.org

Table 5. Experimental effects of ablation on each module of U2net-MT.

The Table 6 shows the comparison results between the extracted features from the second and third classification groups and the original images used directly for tongue classification. For the peeling and greasy tongue coatings, which have significant texture differences in grayscale images, LBP feature extraction significantly improved classification performance by filtering out much irrelevant noise. In the third group, related to the moist-dry classification, the proportion of relevant features in the images was too low for the model to extract effective information for classification at various levels. The experiments demonstrated that wavelet transform, which provides time-frequency localization, could accurately capture subtle texture features in tongue images. The wavelet features effectively reflect the moist-dry correlation, solving the challenging moist-dry classification problem. The fusion of original image features added complementary detail features, further improving classification performance. Introducing CBAM channel and spatial attention mechanisms allowed the model to focus more on key features, enhancing classification accuracy.

Table 6
www.frontiersin.org

Table 6. Results of ablation experiments on the strategy of using LBP and wavelet features in tongue image classification (%).

5.3 Conclusion

In this study, a complete tongue image analysis system was successfully developed, combining modern deep learning techniques with traditional Chinese medicine tongue diagnosis to improve the accuracy of tongue segmentation and coating assessment. Specifically, the semi-supervised learning algorithm based on the U2Net model significantly improved the quality of image segmentation. In addition, the color correction module ensured the accuracy and consistency of image data, and wavelet features were integrated for tongue diagnosis analysis. Experimental results demonstrated the system’s outstanding performance in feature extraction and classification of tongue images. Furthermore, the color correction strategy effectively resolved color deviations caused by device differences and environmental variations, providing a more reliable foundation for tongue image analysis. The integration of wavelet features also effectively addressed the challenging problem of moist-dry classification. In future work, the relationship between patients’ tongue characteristics and endoscopic examination results will be analyzed to explore their correlation.

This work not only realizes automated tongue analysis, but also provides real-time feedback of the analysis results, reducing the time and effort required for manual diagnosis. This is important for improving diagnosis and treatment efficiency and reducing the workload of medical staff, especially in large-scale patient management and telemedicine.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by the Medical Ethics Committee of Sichuan Second Hospital of Traditional Chinese Medicine. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and institutional requirements.

Author contributions

TC: Conceptualization, Investigation, Writing–original draft. YC: Investigation, Methodology, Software, Writing–original draft, Writing–review and editing. ZZ: Formal Analysis, Funding acquisition, Project administration, Resources, Writing–original draft. YZ: Investigation, Project administration, Software, Validation, Writing–original draft. LH: Formal Analysis, Investigation, Methodology, Software, Writing–review and editing. JZ: Supervision, Validation, Writing–review and editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This study was supported by the Sichuan Science and Technology Program (2023YFS0327), (2024YFFK0044) and (2024YFFK0089).

Acknowledgments

For this, I’m extremely grateful. I would like to express our sincere gratitude to the School of Biomedical Engineering at Sichuan University for providing valuable academic resources for the research materials referenced in this paper. My thanks also go to the Second Hospital of Sichuan Provincial Traditional Chinese Medicine for their constructive suggestions and assistance in the collection and annotation of the data used in this study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphys.2025.1559389/full#supplementary-material

References

Alabi R. O., Almangush A., Elmusrati M., Mäkitie A. A. (2022). Deep machine learning for oral cancer: from precise diagnosis to precision medicine. Front. Oral Health 2, 794248. doi:10.3389/froh.2021.794248

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen J., Du J., Feng C., Li J. (2023). “Automatic classfication of tongue color based on image processing,” in 2023 international seminar on computer science and engineering technology (SCSET) (New York, NY, USA), 257–261.

Google Scholar

Chen L.-C., Zhu Y., Papandreou G., Schroff F., Adam H. (2018). Encoder-decoder with atrous separable convolution for semantic image segmentation, 833, 851. doi:10.1007/978-3-030-01234-2_49

CrossRef Full Text | Google Scholar

Dosovitskiy A., Beyer L., Kolesnikov A., Weissenborn D., Zhai X., Unterthiner T., et al. (2021). An image is worth 16x16 words: transformers for image recognition at scale. ICLR.

Google Scholar

Gao F., Hu M., Zhong M.-E., Feng S., Tian X., Meng X., et al. (2022). Segmentation only uses sparse annotations: unified weakly and semi-supervised learning in medical images. Med. Image Anal. 80, 102515. doi:10.1016/j.media.2022.102515

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao S., Guo N., Mao D. (2021). Lsm-sec: tongue segmentation by the level set model with symmetry and edge constraints. Comput. Intell. Neurosci. 2021, 6370526. doi:10.1155/2021/6370526

PubMed Abstract | CrossRef Full Text | Google Scholar

Howard A. G., Zhu M., Chen B., Kalenichenko D., Wang W., Weyand T., et al. (2017). Mobilenets: efficient convolutional neural networks for mobile vision applications.

Google Scholar

Kang G., Hao Y., Wang Y., Cao G., Ma Z., Xia C. (2024). “A two-stage approach for mobile-acquired tongue image with yolov5 and la-unet,” in 2024 IEEE 4th international Conference on electronic technology, Communication and information (ICETCI). 454–60. 2024 IEEE 4th international conference on electronic technology, communication and information (ICETCI), 2024 (China: Changchun).

Google Scholar

Long J., Shelhamer E., Darrell T. (2014). Fully convolutional networks for semantic segmentation.

Google Scholar

Massa F., Girshick R. (2018). Maskrcnn-benchmark: fast, modular reference implementation of instance segmentation and object detection algorithms in PyTorch. Available online at: https://github.com/facebookresearch/maskrcnn-benchmark.

Google Scholar

Ojala T., Pietikainen M., Harwood D. (1994). “Performance evaluation of texture measures with classification based on kullback discrimination of distributions,” in Proceedings of the 12TH iapr international conference on pattern recognition - conference a: computer VISION and image processing (int asSOC pattern recognit; ieee, comp SOC; informat proc asSOC Israel), international conference on pattern recognition, 582–585. doi:10.1109/icpr.1994.576366

CrossRef Full Text | Google Scholar

Prisilla A. A., Guo Y. L., Jan Y.-K., Lin C.-Y., Lin F.-Y., Liau B.-Y., et al. (2023). An approach to the diagnosis of lumbar disc herniation using deep learning models. Front. Bioeng. Biotechnol. 11, 1247112. doi:10.3389/fbioe.2023.1247112

PubMed Abstract | CrossRef Full Text | Google Scholar

Qin X., Zhang Z., Huang C., Dehghan M., Zaiane O., Jagersand M. (2020). U2-net: going deeper with nested u-structure for salient object detection [arxiv]. arXiv, 15.

Google Scholar

Ronneberger O., Fischer P., Brox T. (2015). U-net: convolutional networks for biomedical image segmentation, 234, 241. doi:10.1007/978-3-319-24574-4_28

CrossRef Full Text | Google Scholar

Stricker M., Orengo M. (1995). “Similarity of color images. In Storage and retrieval for image and video databases iii,” in (SOC IMAGING SCI and TECHNOL; SOC PHOTO OPT INSTRUMENTAT ENGINEERS), vol. 2410 of PROCEEDINGS OF THE SOCIETY OF PHOTO-OPTICAL INSTRUMENTATION ENGINEERS (SPIE). Conference on storage and retrieval for image and video databases III. Editors W. Niblack, and R. Jain, 381–392. SAN JOSE, CA, FEB 09-10, 1995.

Google Scholar

Syed S. R., M A S. D. (2023). A diagnosis model for brain atrophy using deep learning and mri of type 2 diabetes mellitus. Front. Neurosci. 17, 1291753. doi:10.3389/fnins.2023.1291753

PubMed Abstract | CrossRef Full Text | Google Scholar

Takikawa T., Acuna D., Jampani V., Fidler S. (2019). “Gated-scnn: Gated shape conns for semantic segmentation,” in ICCV.

Google Scholar

Tarvainen A., Valpola H. (2017). “Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results,” in Advances in neural information processing systems. Editors I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathanet al. (Curran Associates, Inc.), 30.

Google Scholar

Woo S., Park J., Lee J.-Y., Kweon I. S. (2018). Cam: convolutional block attention module, 3, 19. doi:10.1007/978-3-030-01234-2_1

CrossRef Full Text | Google Scholar

Xin G., Zhu L., Liang H., Ding C. (2021). “A fast tongue image color correction method based on gray world method,” in Artificial intelligence and security: 7th international conference, ICAIS 2021. Lecture notes in computer science, information systems and applications, incl. Internet/web, and HCI (12737). Editors X. Sun, X. Zhang, Z. Xia, and E. Bertino (Dublin, Ireland), 724–735.

Google Scholar

Yan B., Zhang S., Su H., Zheng H. (2021). “Tccgan: a stacked generative adversarial network for clinical tongue images color correction,” in Icdsp 2021: 2021 5th international Conference on digital signal processing. 34–9. ICDSP 2021: 2021 5th international conference on digital signal processing, 26-28 feb. 2021 (Chengdu, China).

Google Scholar

Yiqin Y. J. B. R. M. H. Z., Chen B., Guo R., Zeng M., Yan H., Xu Z., et al. (2022). Tongue image texture classification based on image inpainting and convolutional neural network. Comput. Math. methods Med. 2022, 6066640. doi:10.1155/2022/6066640

PubMed Abstract | CrossRef Full Text | Google Scholar

Yuan L., Yang L., Zhang S., Xu Z., Qin J., Shi Y., et al. (2023). Development of a tongue image-based machine learning tool for the diagnosis of gastric cancer: a prospective multicentre clinical cohort study. ECLINICALMEDICINE 57, 101834. doi:10.1016/j.eclinm.2023.101834

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang K., Jiang J., Zhang H., Xiong Z., Zhu M., Tao Q. (2023). “Automatic classification of tongue coating thickness based on image processing technology,” in 2023 IEEE 5th eurasia conference on IOT, communication and engineering (ECICE). Editor T.-H. Meen (Yunlin, Taiwan), 368–370.

Google Scholar

Zhang R., Lai P., Wan X., Fan D.-J., Gao F., Wu X.-J., et al. (2022a). “Lesion-aware dynamic kernel for polyp segmentation,” in Medical image computing and computer assisted intervention – miccai 2022. Editors L. Wang, Q. Dou, P. T. Fletcher, S. Speidel, and S. Li (Cham: Springer Nature Switzerland), 99–109.

CrossRef Full Text | Google Scholar

Zhang X., Bian H., Cai Y., Zhang K., Li H. (2022b). An improved tongue image segmentation algorithm based on deeplabv3+framework. IET IMAGE Process. 16, 1473–1485. doi:10.1049/ipr2.12425

CrossRef Full Text | Google Scholar

Zhao X., Fang C., Fan D.-J., Lin X., Gao F., Li G. (2022). “Cross-level contrastive learning and consistency constraint for semi-supervised medical image segmentation,” in 2022 IEEE 19th international symposium on biomedical imaging (ISBI), 1–5. doi:10.1109/ISBI52829.2022.9761710

CrossRef Full Text | Google Scholar

Zhong M.-E., Duan X., Ni-jia ti M.-y.-d.-l., Qi H., Xu D., Cai D., et al. (2022). Ct-based radiogenomic analysis dissects intratumor heterogeneity and predicts prognosis of colorectal cancer: a multi-institutional retrospective study. J. Transl. Med. 20, 574. doi:10.1186/s12967-022-03788-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou X., Wang H., Feng C., Xu R., He Y., Li L., et al. (2022). Emerging applications of deep learning in bone tumors: current advances and challenges. Front. Oncol. 12, 908873. doi:10.3389/fonc.2022.908873

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou Z., Siddiquee M. M. R., Tajbakhsh N., Liang J. (2020). Unet++: redesigning skip connections to exploit multiscale features in image segmentation.

Google Scholar

Keywords: tongue images, image segmentation, image classification, tongue diagnosis, neural networks, color correction

Citation: Chen T, Chen Y, Zhou Z, Zhu Y, He L and Zhang J (2025) Deep learning-based automated tongue analysis system for assisted Chinese medicine diagnosis. Front. Physiol. 16:1559389. doi: 10.3389/fphys.2025.1559389

Received: 12 January 2025; Accepted: 27 February 2025;
Published: 28 April 2025.

Edited by:

Feng Gao, The Sixth Affiliated Hospital of Sun Yat-sen University, China

Reviewed by:

Huiqi Li, Beijing Institute of Technology, China
Jun Cheng, A∗STAR Graduate Academy (A∗STAR), Singapore

Copyright © 2025 Chen, Chen, Zhou, Zhu, He and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jing Zhang, amluZ196aGFuZ0BzY3UuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.